12026 lines
299 KiB
C
12026 lines
299 KiB
C
/* tc-ia64.c -- Assembler for the HP/Intel IA-64 architecture.
|
|
Copyright (C) 1998-2018 Free Software Foundation, Inc.
|
|
Contributed by David Mosberger-Tang <davidm@hpl.hp.com>
|
|
|
|
This file is part of GAS, the GNU Assembler.
|
|
|
|
GAS is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
any later version.
|
|
|
|
GAS is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GAS; see the file COPYING. If not, write to
|
|
the Free Software Foundation, 51 Franklin Street - Fifth Floor,
|
|
Boston, MA 02110-1301, USA. */
|
|
|
|
/*
|
|
TODO:
|
|
|
|
- optional operands
|
|
- directives:
|
|
.eb
|
|
.estate
|
|
.lb
|
|
.popsection
|
|
.previous
|
|
.psr
|
|
.pushsection
|
|
- labels are wrong if automatic alignment is introduced
|
|
(e.g., checkout the second real10 definition in test-data.s)
|
|
- DV-related stuff:
|
|
<reg>.safe_across_calls and any other DV-related directives I don't
|
|
have documentation for.
|
|
verify mod-sched-brs reads/writes are checked/marked (and other
|
|
notes)
|
|
|
|
*/
|
|
|
|
#include "as.h"
|
|
#include "safe-ctype.h"
|
|
#include "dwarf2dbg.h"
|
|
#include "subsegs.h"
|
|
|
|
#include "opcode/ia64.h"
|
|
|
|
#include "elf/ia64.h"
|
|
#include "bfdver.h"
|
|
#include <time.h>
|
|
|
|
#ifdef HAVE_LIMITS_H
|
|
#include <limits.h>
|
|
#endif
|
|
|
|
#define NELEMS(a) ((int) (sizeof (a)/sizeof ((a)[0])))
|
|
|
|
/* Some systems define MIN in, e.g., param.h. */
|
|
#undef MIN
|
|
#define MIN(a,b) ((a) < (b) ? (a) : (b))
|
|
|
|
#define NUM_SLOTS 4
|
|
#define PREV_SLOT md.slot[(md.curr_slot + NUM_SLOTS - 1) % NUM_SLOTS]
|
|
#define CURR_SLOT md.slot[md.curr_slot]
|
|
|
|
#define O_pseudo_fixup (O_max + 1)
|
|
|
|
enum special_section
|
|
{
|
|
/* IA-64 ABI section pseudo-ops. */
|
|
SPECIAL_SECTION_BSS = 0,
|
|
SPECIAL_SECTION_SBSS,
|
|
SPECIAL_SECTION_SDATA,
|
|
SPECIAL_SECTION_RODATA,
|
|
SPECIAL_SECTION_COMMENT,
|
|
SPECIAL_SECTION_UNWIND,
|
|
SPECIAL_SECTION_UNWIND_INFO,
|
|
/* HPUX specific section pseudo-ops. */
|
|
SPECIAL_SECTION_INIT_ARRAY,
|
|
SPECIAL_SECTION_FINI_ARRAY,
|
|
};
|
|
|
|
enum reloc_func
|
|
{
|
|
FUNC_DTP_MODULE,
|
|
FUNC_DTP_RELATIVE,
|
|
FUNC_FPTR_RELATIVE,
|
|
FUNC_GP_RELATIVE,
|
|
FUNC_LT_RELATIVE,
|
|
FUNC_LT_RELATIVE_X,
|
|
FUNC_PC_RELATIVE,
|
|
FUNC_PLT_RELATIVE,
|
|
FUNC_SEC_RELATIVE,
|
|
FUNC_SEG_RELATIVE,
|
|
FUNC_TP_RELATIVE,
|
|
FUNC_LTV_RELATIVE,
|
|
FUNC_LT_FPTR_RELATIVE,
|
|
FUNC_LT_DTP_MODULE,
|
|
FUNC_LT_DTP_RELATIVE,
|
|
FUNC_LT_TP_RELATIVE,
|
|
FUNC_IPLT_RELOC,
|
|
#ifdef TE_VMS
|
|
FUNC_SLOTCOUNT_RELOC,
|
|
#endif
|
|
};
|
|
|
|
enum reg_symbol
|
|
{
|
|
REG_GR = 0,
|
|
REG_FR = (REG_GR + 128),
|
|
REG_AR = (REG_FR + 128),
|
|
REG_CR = (REG_AR + 128),
|
|
REG_DAHR = (REG_CR + 128),
|
|
REG_P = (REG_DAHR + 8),
|
|
REG_BR = (REG_P + 64),
|
|
REG_IP = (REG_BR + 8),
|
|
REG_CFM,
|
|
REG_PR,
|
|
REG_PR_ROT,
|
|
REG_PSR,
|
|
REG_PSR_L,
|
|
REG_PSR_UM,
|
|
/* The following are pseudo-registers for use by gas only. */
|
|
IND_CPUID,
|
|
IND_DBR,
|
|
IND_DTR,
|
|
IND_ITR,
|
|
IND_IBR,
|
|
IND_MSR,
|
|
IND_PKR,
|
|
IND_PMC,
|
|
IND_PMD,
|
|
IND_DAHR,
|
|
IND_RR,
|
|
/* The following pseudo-registers are used for unwind directives only: */
|
|
REG_PSP,
|
|
REG_PRIUNAT,
|
|
REG_NUM
|
|
};
|
|
|
|
enum dynreg_type
|
|
{
|
|
DYNREG_GR = 0, /* dynamic general purpose register */
|
|
DYNREG_FR, /* dynamic floating point register */
|
|
DYNREG_PR, /* dynamic predicate register */
|
|
DYNREG_NUM_TYPES
|
|
};
|
|
|
|
enum operand_match_result
|
|
{
|
|
OPERAND_MATCH,
|
|
OPERAND_OUT_OF_RANGE,
|
|
OPERAND_MISMATCH
|
|
};
|
|
|
|
/* On the ia64, we can't know the address of a text label until the
|
|
instructions are packed into a bundle. To handle this, we keep
|
|
track of the list of labels that appear in front of each
|
|
instruction. */
|
|
struct label_fix
|
|
{
|
|
struct label_fix *next;
|
|
struct symbol *sym;
|
|
bfd_boolean dw2_mark_labels;
|
|
};
|
|
|
|
#ifdef TE_VMS
|
|
/* An internally used relocation. */
|
|
#define DUMMY_RELOC_IA64_SLOTCOUNT (BFD_RELOC_UNUSED + 1)
|
|
#endif
|
|
|
|
/* This is the endianness of the current section. */
|
|
extern int target_big_endian;
|
|
|
|
/* This is the default endianness. */
|
|
static int default_big_endian = TARGET_BYTES_BIG_ENDIAN;
|
|
|
|
void (*ia64_number_to_chars) (char *, valueT, int);
|
|
|
|
static void ia64_float_to_chars_bigendian (char *, LITTLENUM_TYPE *, int);
|
|
static void ia64_float_to_chars_littleendian (char *, LITTLENUM_TYPE *, int);
|
|
|
|
static void (*ia64_float_to_chars) (char *, LITTLENUM_TYPE *, int);
|
|
|
|
static struct hash_control *alias_hash;
|
|
static struct hash_control *alias_name_hash;
|
|
static struct hash_control *secalias_hash;
|
|
static struct hash_control *secalias_name_hash;
|
|
|
|
/* List of chars besides those in app.c:symbol_chars that can start an
|
|
operand. Used to prevent the scrubber eating vital white-space. */
|
|
const char ia64_symbol_chars[] = "@?";
|
|
|
|
/* Characters which always start a comment. */
|
|
const char comment_chars[] = "";
|
|
|
|
/* Characters which start a comment at the beginning of a line. */
|
|
const char line_comment_chars[] = "#";
|
|
|
|
/* Characters which may be used to separate multiple commands on a
|
|
single line. */
|
|
const char line_separator_chars[] = ";{}";
|
|
|
|
/* Characters which are used to indicate an exponent in a floating
|
|
point number. */
|
|
const char EXP_CHARS[] = "eE";
|
|
|
|
/* Characters which mean that a number is a floating point constant,
|
|
as in 0d1.0. */
|
|
const char FLT_CHARS[] = "rRsSfFdDxXpP";
|
|
|
|
/* ia64-specific option processing: */
|
|
|
|
const char *md_shortopts = "m:N:x::";
|
|
|
|
struct option md_longopts[] =
|
|
{
|
|
#define OPTION_MCONSTANT_GP (OPTION_MD_BASE + 1)
|
|
{"mconstant-gp", no_argument, NULL, OPTION_MCONSTANT_GP},
|
|
#define OPTION_MAUTO_PIC (OPTION_MD_BASE + 2)
|
|
{"mauto-pic", no_argument, NULL, OPTION_MAUTO_PIC}
|
|
};
|
|
|
|
size_t md_longopts_size = sizeof (md_longopts);
|
|
|
|
static struct
|
|
{
|
|
struct hash_control *pseudo_hash; /* pseudo opcode hash table */
|
|
struct hash_control *reg_hash; /* register name hash table */
|
|
struct hash_control *dynreg_hash; /* dynamic register hash table */
|
|
struct hash_control *const_hash; /* constant hash table */
|
|
struct hash_control *entry_hash; /* code entry hint hash table */
|
|
|
|
/* If X_op is != O_absent, the register name for the instruction's
|
|
qualifying predicate. If NULL, p0 is assumed for instructions
|
|
that are predictable. */
|
|
expressionS qp;
|
|
|
|
/* Optimize for which CPU. */
|
|
enum
|
|
{
|
|
itanium1,
|
|
itanium2
|
|
} tune;
|
|
|
|
/* What to do when hint.b is used. */
|
|
enum
|
|
{
|
|
hint_b_error,
|
|
hint_b_warning,
|
|
hint_b_ok
|
|
} hint_b;
|
|
|
|
unsigned int
|
|
manual_bundling : 1,
|
|
debug_dv: 1,
|
|
detect_dv: 1,
|
|
explicit_mode : 1, /* which mode we're in */
|
|
default_explicit_mode : 1, /* which mode is the default */
|
|
mode_explicitly_set : 1, /* was the current mode explicitly set? */
|
|
auto_align : 1,
|
|
keep_pending_output : 1;
|
|
|
|
/* What to do when something is wrong with unwind directives. */
|
|
enum
|
|
{
|
|
unwind_check_warning,
|
|
unwind_check_error
|
|
} unwind_check;
|
|
|
|
/* Each bundle consists of up to three instructions. We keep
|
|
track of four most recent instructions so we can correctly set
|
|
the end_of_insn_group for the last instruction in a bundle. */
|
|
int curr_slot;
|
|
int num_slots_in_use;
|
|
struct slot
|
|
{
|
|
unsigned int
|
|
end_of_insn_group : 1,
|
|
manual_bundling_on : 1,
|
|
manual_bundling_off : 1,
|
|
loc_directive_seen : 1;
|
|
signed char user_template; /* user-selected template, if any */
|
|
unsigned char qp_regno; /* qualifying predicate */
|
|
/* This duplicates a good fraction of "struct fix" but we
|
|
can't use a "struct fix" instead since we can't call
|
|
fix_new_exp() until we know the address of the instruction. */
|
|
int num_fixups;
|
|
struct insn_fix
|
|
{
|
|
bfd_reloc_code_real_type code;
|
|
enum ia64_opnd opnd; /* type of operand in need of fix */
|
|
unsigned int is_pcrel : 1; /* is operand pc-relative? */
|
|
expressionS expr; /* the value to be inserted */
|
|
}
|
|
fixup[2]; /* at most two fixups per insn */
|
|
struct ia64_opcode *idesc;
|
|
struct label_fix *label_fixups;
|
|
struct label_fix *tag_fixups;
|
|
struct unw_rec_list *unwind_record; /* Unwind directive. */
|
|
expressionS opnd[6];
|
|
const char *src_file;
|
|
unsigned int src_line;
|
|
struct dwarf2_line_info debug_line;
|
|
}
|
|
slot[NUM_SLOTS];
|
|
|
|
segT last_text_seg;
|
|
|
|
struct dynreg
|
|
{
|
|
struct dynreg *next; /* next dynamic register */
|
|
const char *name;
|
|
unsigned short base; /* the base register number */
|
|
unsigned short num_regs; /* # of registers in this set */
|
|
}
|
|
*dynreg[DYNREG_NUM_TYPES], in, loc, out, rot;
|
|
|
|
flagword flags; /* ELF-header flags */
|
|
|
|
struct mem_offset {
|
|
unsigned hint:1; /* is this hint currently valid? */
|
|
bfd_vma offset; /* mem.offset offset */
|
|
bfd_vma base; /* mem.offset base */
|
|
} mem_offset;
|
|
|
|
int path; /* number of alt. entry points seen */
|
|
const char **entry_labels; /* labels of all alternate paths in
|
|
the current DV-checking block. */
|
|
int maxpaths; /* size currently allocated for
|
|
entry_labels */
|
|
|
|
int pointer_size; /* size in bytes of a pointer */
|
|
int pointer_size_shift; /* shift size of a pointer for alignment */
|
|
|
|
symbolS *indregsym[IND_RR - IND_CPUID + 1];
|
|
}
|
|
md;
|
|
|
|
/* These are not const, because they are modified to MMI for non-itanium1
|
|
targets below. */
|
|
/* MFI bundle of nops. */
|
|
static unsigned char le_nop[16] =
|
|
{
|
|
0x0c, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00,
|
|
0x00, 0x02, 0x00, 0x00, 0x00, 0x00, 0x04, 0x00
|
|
};
|
|
/* MFI bundle of nops with stop-bit. */
|
|
static unsigned char le_nop_stop[16] =
|
|
{
|
|
0x0d, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00,
|
|
0x00, 0x02, 0x00, 0x00, 0x00, 0x00, 0x04, 0x00
|
|
};
|
|
|
|
/* application registers: */
|
|
|
|
#define AR_K0 0
|
|
#define AR_K7 7
|
|
#define AR_RSC 16
|
|
#define AR_BSP 17
|
|
#define AR_BSPSTORE 18
|
|
#define AR_RNAT 19
|
|
#define AR_FCR 21
|
|
#define AR_EFLAG 24
|
|
#define AR_CSD 25
|
|
#define AR_SSD 26
|
|
#define AR_CFLG 27
|
|
#define AR_FSR 28
|
|
#define AR_FIR 29
|
|
#define AR_FDR 30
|
|
#define AR_CCV 32
|
|
#define AR_UNAT 36
|
|
#define AR_FPSR 40
|
|
#define AR_ITC 44
|
|
#define AR_RUC 45
|
|
#define AR_PFS 64
|
|
#define AR_LC 65
|
|
#define AR_EC 66
|
|
|
|
static const struct
|
|
{
|
|
const char *name;
|
|
unsigned int regnum;
|
|
}
|
|
ar[] =
|
|
{
|
|
{"ar.k0", AR_K0}, {"ar.k1", AR_K0 + 1},
|
|
{"ar.k2", AR_K0 + 2}, {"ar.k3", AR_K0 + 3},
|
|
{"ar.k4", AR_K0 + 4}, {"ar.k5", AR_K0 + 5},
|
|
{"ar.k6", AR_K0 + 6}, {"ar.k7", AR_K7},
|
|
{"ar.rsc", AR_RSC}, {"ar.bsp", AR_BSP},
|
|
{"ar.bspstore", AR_BSPSTORE}, {"ar.rnat", AR_RNAT},
|
|
{"ar.fcr", AR_FCR}, {"ar.eflag", AR_EFLAG},
|
|
{"ar.csd", AR_CSD}, {"ar.ssd", AR_SSD},
|
|
{"ar.cflg", AR_CFLG}, {"ar.fsr", AR_FSR},
|
|
{"ar.fir", AR_FIR}, {"ar.fdr", AR_FDR},
|
|
{"ar.ccv", AR_CCV}, {"ar.unat", AR_UNAT},
|
|
{"ar.fpsr", AR_FPSR}, {"ar.itc", AR_ITC},
|
|
{"ar.ruc", AR_RUC}, {"ar.pfs", AR_PFS},
|
|
{"ar.lc", AR_LC}, {"ar.ec", AR_EC},
|
|
};
|
|
|
|
/* control registers: */
|
|
|
|
#define CR_DCR 0
|
|
#define CR_ITM 1
|
|
#define CR_IVA 2
|
|
#define CR_PTA 8
|
|
#define CR_GPTA 9
|
|
#define CR_IPSR 16
|
|
#define CR_ISR 17
|
|
#define CR_IIP 19
|
|
#define CR_IFA 20
|
|
#define CR_ITIR 21
|
|
#define CR_IIPA 22
|
|
#define CR_IFS 23
|
|
#define CR_IIM 24
|
|
#define CR_IHA 25
|
|
#define CR_IIB0 26
|
|
#define CR_IIB1 27
|
|
#define CR_LID 64
|
|
#define CR_IVR 65
|
|
#define CR_TPR 66
|
|
#define CR_EOI 67
|
|
#define CR_IRR0 68
|
|
#define CR_IRR3 71
|
|
#define CR_ITV 72
|
|
#define CR_PMV 73
|
|
#define CR_CMCV 74
|
|
#define CR_LRR0 80
|
|
#define CR_LRR1 81
|
|
|
|
static const struct
|
|
{
|
|
const char *name;
|
|
unsigned int regnum;
|
|
}
|
|
cr[] =
|
|
{
|
|
{"cr.dcr", CR_DCR},
|
|
{"cr.itm", CR_ITM},
|
|
{"cr.iva", CR_IVA},
|
|
{"cr.pta", CR_PTA},
|
|
{"cr.gpta", CR_GPTA},
|
|
{"cr.ipsr", CR_IPSR},
|
|
{"cr.isr", CR_ISR},
|
|
{"cr.iip", CR_IIP},
|
|
{"cr.ifa", CR_IFA},
|
|
{"cr.itir", CR_ITIR},
|
|
{"cr.iipa", CR_IIPA},
|
|
{"cr.ifs", CR_IFS},
|
|
{"cr.iim", CR_IIM},
|
|
{"cr.iha", CR_IHA},
|
|
{"cr.iib0", CR_IIB0},
|
|
{"cr.iib1", CR_IIB1},
|
|
{"cr.lid", CR_LID},
|
|
{"cr.ivr", CR_IVR},
|
|
{"cr.tpr", CR_TPR},
|
|
{"cr.eoi", CR_EOI},
|
|
{"cr.irr0", CR_IRR0},
|
|
{"cr.irr1", CR_IRR0 + 1},
|
|
{"cr.irr2", CR_IRR0 + 2},
|
|
{"cr.irr3", CR_IRR3},
|
|
{"cr.itv", CR_ITV},
|
|
{"cr.pmv", CR_PMV},
|
|
{"cr.cmcv", CR_CMCV},
|
|
{"cr.lrr0", CR_LRR0},
|
|
{"cr.lrr1", CR_LRR1}
|
|
};
|
|
|
|
#define PSR_MFL 4
|
|
#define PSR_IC 13
|
|
#define PSR_DFL 18
|
|
#define PSR_CPL 32
|
|
|
|
static const struct const_desc
|
|
{
|
|
const char *name;
|
|
valueT value;
|
|
}
|
|
const_bits[] =
|
|
{
|
|
/* PSR constant masks: */
|
|
|
|
/* 0: reserved */
|
|
{"psr.be", ((valueT) 1) << 1},
|
|
{"psr.up", ((valueT) 1) << 2},
|
|
{"psr.ac", ((valueT) 1) << 3},
|
|
{"psr.mfl", ((valueT) 1) << 4},
|
|
{"psr.mfh", ((valueT) 1) << 5},
|
|
/* 6-12: reserved */
|
|
{"psr.ic", ((valueT) 1) << 13},
|
|
{"psr.i", ((valueT) 1) << 14},
|
|
{"psr.pk", ((valueT) 1) << 15},
|
|
/* 16: reserved */
|
|
{"psr.dt", ((valueT) 1) << 17},
|
|
{"psr.dfl", ((valueT) 1) << 18},
|
|
{"psr.dfh", ((valueT) 1) << 19},
|
|
{"psr.sp", ((valueT) 1) << 20},
|
|
{"psr.pp", ((valueT) 1) << 21},
|
|
{"psr.di", ((valueT) 1) << 22},
|
|
{"psr.si", ((valueT) 1) << 23},
|
|
{"psr.db", ((valueT) 1) << 24},
|
|
{"psr.lp", ((valueT) 1) << 25},
|
|
{"psr.tb", ((valueT) 1) << 26},
|
|
{"psr.rt", ((valueT) 1) << 27},
|
|
/* 28-31: reserved */
|
|
/* 32-33: cpl (current privilege level) */
|
|
{"psr.is", ((valueT) 1) << 34},
|
|
{"psr.mc", ((valueT) 1) << 35},
|
|
{"psr.it", ((valueT) 1) << 36},
|
|
{"psr.id", ((valueT) 1) << 37},
|
|
{"psr.da", ((valueT) 1) << 38},
|
|
{"psr.dd", ((valueT) 1) << 39},
|
|
{"psr.ss", ((valueT) 1) << 40},
|
|
/* 41-42: ri (restart instruction) */
|
|
{"psr.ed", ((valueT) 1) << 43},
|
|
{"psr.bn", ((valueT) 1) << 44},
|
|
};
|
|
|
|
/* indirect register-sets/memory: */
|
|
|
|
static const struct
|
|
{
|
|
const char *name;
|
|
unsigned int regnum;
|
|
}
|
|
indirect_reg[] =
|
|
{
|
|
{ "CPUID", IND_CPUID },
|
|
{ "cpuid", IND_CPUID },
|
|
{ "dbr", IND_DBR },
|
|
{ "dtr", IND_DTR },
|
|
{ "itr", IND_ITR },
|
|
{ "ibr", IND_IBR },
|
|
{ "msr", IND_MSR },
|
|
{ "pkr", IND_PKR },
|
|
{ "pmc", IND_PMC },
|
|
{ "pmd", IND_PMD },
|
|
{ "dahr", IND_DAHR },
|
|
{ "rr", IND_RR },
|
|
};
|
|
|
|
/* Pseudo functions used to indicate relocation types (these functions
|
|
start with an at sign (@). */
|
|
static struct
|
|
{
|
|
const char *name;
|
|
enum pseudo_type
|
|
{
|
|
PSEUDO_FUNC_NONE,
|
|
PSEUDO_FUNC_RELOC,
|
|
PSEUDO_FUNC_CONST,
|
|
PSEUDO_FUNC_REG,
|
|
PSEUDO_FUNC_FLOAT
|
|
}
|
|
type;
|
|
union
|
|
{
|
|
unsigned long ival;
|
|
symbolS *sym;
|
|
}
|
|
u;
|
|
}
|
|
pseudo_func[] =
|
|
{
|
|
/* reloc pseudo functions (these must come first!): */
|
|
{ "dtpmod", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "dtprel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "fptr", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "gprel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "ltoff", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "ltoffx", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "pcrel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "pltoff", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "secrel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "segrel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "tprel", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ "ltv", PSEUDO_FUNC_RELOC, { 0 } },
|
|
{ NULL, 0, { 0 } }, /* placeholder for FUNC_LT_FPTR_RELATIVE */
|
|
{ NULL, 0, { 0 } }, /* placeholder for FUNC_LT_DTP_MODULE */
|
|
{ NULL, 0, { 0 } }, /* placeholder for FUNC_LT_DTP_RELATIVE */
|
|
{ NULL, 0, { 0 } }, /* placeholder for FUNC_LT_TP_RELATIVE */
|
|
{ "iplt", PSEUDO_FUNC_RELOC, { 0 } },
|
|
#ifdef TE_VMS
|
|
{ "slotcount", PSEUDO_FUNC_RELOC, { 0 } },
|
|
#endif
|
|
|
|
/* mbtype4 constants: */
|
|
{ "alt", PSEUDO_FUNC_CONST, { 0xa } },
|
|
{ "brcst", PSEUDO_FUNC_CONST, { 0x0 } },
|
|
{ "mix", PSEUDO_FUNC_CONST, { 0x8 } },
|
|
{ "rev", PSEUDO_FUNC_CONST, { 0xb } },
|
|
{ "shuf", PSEUDO_FUNC_CONST, { 0x9 } },
|
|
|
|
/* fclass constants: */
|
|
{ "nat", PSEUDO_FUNC_CONST, { 0x100 } },
|
|
{ "qnan", PSEUDO_FUNC_CONST, { 0x080 } },
|
|
{ "snan", PSEUDO_FUNC_CONST, { 0x040 } },
|
|
{ "pos", PSEUDO_FUNC_CONST, { 0x001 } },
|
|
{ "neg", PSEUDO_FUNC_CONST, { 0x002 } },
|
|
{ "zero", PSEUDO_FUNC_CONST, { 0x004 } },
|
|
{ "unorm", PSEUDO_FUNC_CONST, { 0x008 } },
|
|
{ "norm", PSEUDO_FUNC_CONST, { 0x010 } },
|
|
{ "inf", PSEUDO_FUNC_CONST, { 0x020 } },
|
|
|
|
{ "natval", PSEUDO_FUNC_CONST, { 0x100 } }, /* old usage */
|
|
|
|
/* hint constants: */
|
|
{ "pause", PSEUDO_FUNC_CONST, { 0x0 } },
|
|
{ "priority", PSEUDO_FUNC_CONST, { 0x1 } },
|
|
|
|
/* tf constants: */
|
|
{ "clz", PSEUDO_FUNC_CONST, { 32 } },
|
|
{ "mpy", PSEUDO_FUNC_CONST, { 33 } },
|
|
{ "datahints", PSEUDO_FUNC_CONST, { 34 } },
|
|
|
|
/* unwind-related constants: */
|
|
{ "svr4", PSEUDO_FUNC_CONST, { ELFOSABI_NONE } },
|
|
{ "hpux", PSEUDO_FUNC_CONST, { ELFOSABI_HPUX } },
|
|
{ "nt", PSEUDO_FUNC_CONST, { 2 } }, /* conflicts w/ELFOSABI_NETBSD */
|
|
{ "linux", PSEUDO_FUNC_CONST, { ELFOSABI_GNU } },
|
|
{ "freebsd", PSEUDO_FUNC_CONST, { ELFOSABI_FREEBSD } },
|
|
{ "openvms", PSEUDO_FUNC_CONST, { ELFOSABI_OPENVMS } },
|
|
{ "nsk", PSEUDO_FUNC_CONST, { ELFOSABI_NSK } },
|
|
|
|
/* unwind-related registers: */
|
|
{ "priunat",PSEUDO_FUNC_REG, { REG_PRIUNAT } }
|
|
};
|
|
|
|
/* 41-bit nop opcodes (one per unit): */
|
|
static const bfd_vma nop[IA64_NUM_UNITS] =
|
|
{
|
|
0x0000000000LL, /* NIL => break 0 */
|
|
0x0008000000LL, /* I-unit nop */
|
|
0x0008000000LL, /* M-unit nop */
|
|
0x4000000000LL, /* B-unit nop */
|
|
0x0008000000LL, /* F-unit nop */
|
|
0x0000000000LL, /* L-"unit" nop immediate */
|
|
0x0008000000LL, /* X-unit nop */
|
|
};
|
|
|
|
/* Can't be `const' as it's passed to input routines (which have the
|
|
habit of setting temporary sentinels. */
|
|
static char special_section_name[][20] =
|
|
{
|
|
{".bss"}, {".sbss"}, {".sdata"}, {".rodata"}, {".comment"},
|
|
{".IA_64.unwind"}, {".IA_64.unwind_info"},
|
|
{".init_array"}, {".fini_array"}
|
|
};
|
|
|
|
/* The best template for a particular sequence of up to three
|
|
instructions: */
|
|
#define N IA64_NUM_TYPES
|
|
static unsigned char best_template[N][N][N];
|
|
#undef N
|
|
|
|
/* Resource dependencies currently in effect */
|
|
static struct rsrc {
|
|
int depind; /* dependency index */
|
|
const struct ia64_dependency *dependency; /* actual dependency */
|
|
unsigned specific:1, /* is this a specific bit/regno? */
|
|
link_to_qp_branch:1; /* will a branch on the same QP clear it?*/
|
|
int index; /* specific regno/bit within dependency */
|
|
int note; /* optional qualifying note (0 if none) */
|
|
#define STATE_NONE 0
|
|
#define STATE_STOP 1
|
|
#define STATE_SRLZ 2
|
|
int insn_srlz; /* current insn serialization state */
|
|
int data_srlz; /* current data serialization state */
|
|
int qp_regno; /* qualifying predicate for this usage */
|
|
const char *file; /* what file marked this dependency */
|
|
unsigned int line; /* what line marked this dependency */
|
|
struct mem_offset mem_offset; /* optional memory offset hint */
|
|
enum { CMP_NONE, CMP_OR, CMP_AND } cmp_type; /* OR or AND compare? */
|
|
int path; /* corresponding code entry index */
|
|
} *regdeps = NULL;
|
|
static int regdepslen = 0;
|
|
static int regdepstotlen = 0;
|
|
static const char *dv_mode[] = { "RAW", "WAW", "WAR" };
|
|
static const char *dv_sem[] = { "none", "implied", "impliedf",
|
|
"data", "instr", "specific", "stop", "other" };
|
|
static const char *dv_cmp_type[] = { "none", "OR", "AND" };
|
|
|
|
/* Current state of PR mutexation */
|
|
static struct qpmutex {
|
|
valueT prmask;
|
|
int path;
|
|
} *qp_mutexes = NULL; /* QP mutex bitmasks */
|
|
static int qp_mutexeslen = 0;
|
|
static int qp_mutexestotlen = 0;
|
|
static valueT qp_safe_across_calls = 0;
|
|
|
|
/* Current state of PR implications */
|
|
static struct qp_imply {
|
|
unsigned p1:6;
|
|
unsigned p2:6;
|
|
unsigned p2_branched:1;
|
|
int path;
|
|
} *qp_implies = NULL;
|
|
static int qp_implieslen = 0;
|
|
static int qp_impliestotlen = 0;
|
|
|
|
/* Keep track of static GR values so that indirect register usage can
|
|
sometimes be tracked. */
|
|
static struct gr {
|
|
unsigned known:1;
|
|
int path;
|
|
valueT value;
|
|
} gr_values[128] = {
|
|
{
|
|
1,
|
|
#ifdef INT_MAX
|
|
INT_MAX,
|
|
#else
|
|
(((1 << (8 * sizeof(gr_values->path) - 2)) - 1) << 1) + 1,
|
|
#endif
|
|
0
|
|
}
|
|
};
|
|
|
|
/* Remember the alignment frag. */
|
|
static fragS *align_frag;
|
|
|
|
/* These are the routines required to output the various types of
|
|
unwind records. */
|
|
|
|
/* A slot_number is a frag address plus the slot index (0-2). We use the
|
|
frag address here so that if there is a section switch in the middle of
|
|
a function, then instructions emitted to a different section are not
|
|
counted. Since there may be more than one frag for a function, this
|
|
means we also need to keep track of which frag this address belongs to
|
|
so we can compute inter-frag distances. This also nicely solves the
|
|
problem with nops emitted for align directives, which can't easily be
|
|
counted, but can easily be derived from frag sizes. */
|
|
|
|
typedef struct unw_rec_list {
|
|
unwind_record r;
|
|
unsigned long slot_number;
|
|
fragS *slot_frag;
|
|
struct unw_rec_list *next;
|
|
} unw_rec_list;
|
|
|
|
#define SLOT_NUM_NOT_SET (unsigned)-1
|
|
|
|
/* Linked list of saved prologue counts. A very poor
|
|
implementation of a map from label numbers to prologue counts. */
|
|
typedef struct label_prologue_count
|
|
{
|
|
struct label_prologue_count *next;
|
|
unsigned long label_number;
|
|
unsigned int prologue_count;
|
|
} label_prologue_count;
|
|
|
|
typedef struct proc_pending
|
|
{
|
|
symbolS *sym;
|
|
struct proc_pending *next;
|
|
} proc_pending;
|
|
|
|
static struct
|
|
{
|
|
/* Maintain a list of unwind entries for the current function. */
|
|
unw_rec_list *list;
|
|
unw_rec_list *tail;
|
|
|
|
/* Any unwind entries that should be attached to the current slot
|
|
that an insn is being constructed for. */
|
|
unw_rec_list *current_entry;
|
|
|
|
/* These are used to create the unwind table entry for this function. */
|
|
proc_pending proc_pending;
|
|
symbolS *info; /* pointer to unwind info */
|
|
symbolS *personality_routine;
|
|
segT saved_text_seg;
|
|
subsegT saved_text_subseg;
|
|
unsigned int force_unwind_entry : 1; /* force generation of unwind entry? */
|
|
|
|
/* TRUE if processing unwind directives in a prologue region. */
|
|
unsigned int prologue : 1;
|
|
unsigned int prologue_mask : 4;
|
|
unsigned int prologue_gr : 7;
|
|
unsigned int body : 1;
|
|
unsigned int insn : 1;
|
|
unsigned int prologue_count; /* number of .prologues seen so far */
|
|
/* Prologue counts at previous .label_state directives. */
|
|
struct label_prologue_count * saved_prologue_counts;
|
|
|
|
/* List of split up .save-s. */
|
|
unw_p_record *pending_saves;
|
|
} unwind;
|
|
|
|
/* The input value is a negated offset from psp, and specifies an address
|
|
psp - offset. The encoded value is psp + 16 - (4 * offset). Thus we
|
|
must add 16 and divide by 4 to get the encoded value. */
|
|
|
|
#define ENCODED_PSP_OFFSET(OFFSET) (((OFFSET) + 16) / 4)
|
|
|
|
typedef void (*vbyte_func) (int, char *, char *);
|
|
|
|
/* Forward declarations: */
|
|
static void dot_alias (int);
|
|
static int parse_operand_and_eval (expressionS *, int);
|
|
static void emit_one_bundle (void);
|
|
static bfd_reloc_code_real_type ia64_gen_real_reloc_type (struct symbol *,
|
|
bfd_reloc_code_real_type);
|
|
static void insn_group_break (int, int, int);
|
|
static void add_qp_mutex (valueT);
|
|
static void add_qp_imply (int, int);
|
|
static void clear_qp_mutex (valueT);
|
|
static void clear_qp_implies (valueT, valueT);
|
|
static void print_dependency (const char *, int);
|
|
static void instruction_serialization (void);
|
|
static void data_serialization (void);
|
|
static void output_R3_format (vbyte_func, unw_record_type, unsigned long);
|
|
static void output_B3_format (vbyte_func, unsigned long, unsigned long);
|
|
static void output_B4_format (vbyte_func, unw_record_type, unsigned long);
|
|
static void free_saved_prologue_counts (void);
|
|
|
|
/* Determine if application register REGNUM resides only in the integer
|
|
unit (as opposed to the memory unit). */
|
|
static int
|
|
ar_is_only_in_integer_unit (int reg)
|
|
{
|
|
reg -= REG_AR;
|
|
return reg >= 64 && reg <= 111;
|
|
}
|
|
|
|
/* Determine if application register REGNUM resides only in the memory
|
|
unit (as opposed to the integer unit). */
|
|
static int
|
|
ar_is_only_in_memory_unit (int reg)
|
|
{
|
|
reg -= REG_AR;
|
|
return reg >= 0 && reg <= 47;
|
|
}
|
|
|
|
/* Switch to section NAME and create section if necessary. It's
|
|
rather ugly that we have to manipulate input_line_pointer but I
|
|
don't see any other way to accomplish the same thing without
|
|
changing obj-elf.c (which may be the Right Thing, in the end). */
|
|
static void
|
|
set_section (char *name)
|
|
{
|
|
char *saved_input_line_pointer;
|
|
|
|
saved_input_line_pointer = input_line_pointer;
|
|
input_line_pointer = name;
|
|
obj_elf_section (0);
|
|
input_line_pointer = saved_input_line_pointer;
|
|
}
|
|
|
|
/* Map 's' to SHF_IA_64_SHORT. */
|
|
|
|
bfd_vma
|
|
ia64_elf_section_letter (int letter, const char **ptr_msg)
|
|
{
|
|
if (letter == 's')
|
|
return SHF_IA_64_SHORT;
|
|
else if (letter == 'o')
|
|
return SHF_LINK_ORDER;
|
|
#ifdef TE_VMS
|
|
else if (letter == 'O')
|
|
return SHF_IA_64_VMS_OVERLAID;
|
|
else if (letter == 'g')
|
|
return SHF_IA_64_VMS_GLOBAL;
|
|
#endif
|
|
|
|
*ptr_msg = _("bad .section directive: want a,o,s,w,x,M,S,G,T in string");
|
|
return -1;
|
|
}
|
|
|
|
/* Map SHF_IA_64_SHORT to SEC_SMALL_DATA. */
|
|
|
|
flagword
|
|
ia64_elf_section_flags (flagword flags,
|
|
bfd_vma attr,
|
|
int type ATTRIBUTE_UNUSED)
|
|
{
|
|
if (attr & SHF_IA_64_SHORT)
|
|
flags |= SEC_SMALL_DATA;
|
|
return flags;
|
|
}
|
|
|
|
int
|
|
ia64_elf_section_type (const char *str, size_t len)
|
|
{
|
|
#define STREQ(s) ((len == sizeof (s) - 1) && (strncmp (str, s, sizeof (s) - 1) == 0))
|
|
|
|
if (STREQ (ELF_STRING_ia64_unwind_info))
|
|
return SHT_PROGBITS;
|
|
|
|
if (STREQ (ELF_STRING_ia64_unwind_info_once))
|
|
return SHT_PROGBITS;
|
|
|
|
if (STREQ (ELF_STRING_ia64_unwind))
|
|
return SHT_IA_64_UNWIND;
|
|
|
|
if (STREQ (ELF_STRING_ia64_unwind_once))
|
|
return SHT_IA_64_UNWIND;
|
|
|
|
if (STREQ ("unwind"))
|
|
return SHT_IA_64_UNWIND;
|
|
|
|
return -1;
|
|
#undef STREQ
|
|
}
|
|
|
|
static unsigned int
|
|
set_regstack (unsigned int ins,
|
|
unsigned int locs,
|
|
unsigned int outs,
|
|
unsigned int rots)
|
|
{
|
|
/* Size of frame. */
|
|
unsigned int sof;
|
|
|
|
sof = ins + locs + outs;
|
|
if (sof > 96)
|
|
{
|
|
as_bad (_("Size of frame exceeds maximum of 96 registers"));
|
|
return 0;
|
|
}
|
|
if (rots > sof)
|
|
{
|
|
as_warn (_("Size of rotating registers exceeds frame size"));
|
|
return 0;
|
|
}
|
|
md.in.base = REG_GR + 32;
|
|
md.loc.base = md.in.base + ins;
|
|
md.out.base = md.loc.base + locs;
|
|
|
|
md.in.num_regs = ins;
|
|
md.loc.num_regs = locs;
|
|
md.out.num_regs = outs;
|
|
md.rot.num_regs = rots;
|
|
return sof;
|
|
}
|
|
|
|
void
|
|
ia64_flush_insns (void)
|
|
{
|
|
struct label_fix *lfix;
|
|
segT saved_seg;
|
|
subsegT saved_subseg;
|
|
unw_rec_list *ptr;
|
|
bfd_boolean mark;
|
|
|
|
if (!md.last_text_seg)
|
|
return;
|
|
|
|
saved_seg = now_seg;
|
|
saved_subseg = now_subseg;
|
|
|
|
subseg_set (md.last_text_seg, 0);
|
|
|
|
while (md.num_slots_in_use > 0)
|
|
emit_one_bundle (); /* force out queued instructions */
|
|
|
|
/* In case there are labels following the last instruction, resolve
|
|
those now. */
|
|
mark = FALSE;
|
|
for (lfix = CURR_SLOT.label_fixups; lfix; lfix = lfix->next)
|
|
{
|
|
symbol_set_value_now (lfix->sym);
|
|
mark |= lfix->dw2_mark_labels;
|
|
}
|
|
if (mark)
|
|
{
|
|
dwarf2_where (&CURR_SLOT.debug_line);
|
|
CURR_SLOT.debug_line.flags |= DWARF2_FLAG_BASIC_BLOCK;
|
|
dwarf2_gen_line_info (frag_now_fix (), &CURR_SLOT.debug_line);
|
|
dwarf2_consume_line_info ();
|
|
}
|
|
CURR_SLOT.label_fixups = 0;
|
|
|
|
for (lfix = CURR_SLOT.tag_fixups; lfix; lfix = lfix->next)
|
|
symbol_set_value_now (lfix->sym);
|
|
CURR_SLOT.tag_fixups = 0;
|
|
|
|
/* In case there are unwind directives following the last instruction,
|
|
resolve those now. We only handle prologue, body, and endp directives
|
|
here. Give an error for others. */
|
|
for (ptr = unwind.current_entry; ptr; ptr = ptr->next)
|
|
{
|
|
switch (ptr->r.type)
|
|
{
|
|
case prologue:
|
|
case prologue_gr:
|
|
case body:
|
|
case endp:
|
|
ptr->slot_number = (unsigned long) frag_more (0);
|
|
ptr->slot_frag = frag_now;
|
|
break;
|
|
|
|
/* Allow any record which doesn't have a "t" field (i.e.,
|
|
doesn't relate to a particular instruction). */
|
|
case unwabi:
|
|
case br_gr:
|
|
case copy_state:
|
|
case fr_mem:
|
|
case frgr_mem:
|
|
case gr_gr:
|
|
case gr_mem:
|
|
case label_state:
|
|
case rp_br:
|
|
case spill_base:
|
|
case spill_mask:
|
|
/* nothing */
|
|
break;
|
|
|
|
default:
|
|
as_bad (_("Unwind directive not followed by an instruction."));
|
|
break;
|
|
}
|
|
}
|
|
unwind.current_entry = NULL;
|
|
|
|
subseg_set (saved_seg, saved_subseg);
|
|
|
|
if (md.qp.X_op == O_register)
|
|
as_bad (_("qualifying predicate not followed by instruction"));
|
|
}
|
|
|
|
void
|
|
ia64_cons_align (int nbytes)
|
|
{
|
|
if (md.auto_align)
|
|
{
|
|
int log;
|
|
for (log = 0; (nbytes & 1) != 1; nbytes >>= 1)
|
|
log++;
|
|
|
|
do_align (log, NULL, 0, 0);
|
|
}
|
|
}
|
|
|
|
#ifdef TE_VMS
|
|
|
|
/* .vms_common section, symbol, size, alignment */
|
|
|
|
static void
|
|
obj_elf_vms_common (int ignore ATTRIBUTE_UNUSED)
|
|
{
|
|
const char *sec_name;
|
|
char *sym_name;
|
|
char c;
|
|
offsetT size;
|
|
offsetT cur_size;
|
|
offsetT temp;
|
|
symbolS *symbolP;
|
|
segT current_seg = now_seg;
|
|
subsegT current_subseg = now_subseg;
|
|
offsetT log_align;
|
|
|
|
/* Section name. */
|
|
sec_name = obj_elf_section_name ();
|
|
if (sec_name == NULL)
|
|
return;
|
|
|
|
/* Symbol name. */
|
|
SKIP_WHITESPACE ();
|
|
if (*input_line_pointer == ',')
|
|
{
|
|
input_line_pointer++;
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("expected ',' after section name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
c = get_symbol_name (&sym_name);
|
|
|
|
if (input_line_pointer == sym_name)
|
|
{
|
|
(void) restore_line_pointer (c);
|
|
as_bad (_("expected symbol name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
symbolP = symbol_find_or_make (sym_name);
|
|
(void) restore_line_pointer (c);
|
|
|
|
if ((S_IS_DEFINED (symbolP) || symbol_equated_p (symbolP))
|
|
&& !S_IS_COMMON (symbolP))
|
|
{
|
|
as_bad (_("Ignoring attempt to re-define symbol"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
/* Symbol size. */
|
|
SKIP_WHITESPACE ();
|
|
if (*input_line_pointer == ',')
|
|
{
|
|
input_line_pointer++;
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("expected ',' after symbol name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
temp = get_absolute_expression ();
|
|
size = temp;
|
|
size &= ((offsetT) 2 << (stdoutput->arch_info->bits_per_address - 1)) - 1;
|
|
if (temp != size)
|
|
{
|
|
as_warn (_("size (%ld) out of range, ignored"), (long) temp);
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
/* Alignment. */
|
|
SKIP_WHITESPACE ();
|
|
if (*input_line_pointer == ',')
|
|
{
|
|
input_line_pointer++;
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("expected ',' after symbol size"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
log_align = get_absolute_expression ();
|
|
|
|
demand_empty_rest_of_line ();
|
|
|
|
obj_elf_change_section
|
|
(sec_name, SHT_NOBITS, 0,
|
|
SHF_ALLOC | SHF_WRITE | SHF_IA_64_VMS_OVERLAID | SHF_IA_64_VMS_GLOBAL,
|
|
0, NULL, 1, 0);
|
|
|
|
S_SET_VALUE (symbolP, 0);
|
|
S_SET_SIZE (symbolP, size);
|
|
S_SET_EXTERNAL (symbolP);
|
|
S_SET_SEGMENT (symbolP, now_seg);
|
|
|
|
symbol_get_bfdsym (symbolP)->flags |= BSF_OBJECT;
|
|
|
|
record_alignment (now_seg, log_align);
|
|
|
|
cur_size = bfd_section_size (stdoutput, now_seg);
|
|
if ((int) size > cur_size)
|
|
{
|
|
char *pfrag
|
|
= frag_var (rs_fill, 1, 1, (relax_substateT)0, NULL,
|
|
(valueT)size - (valueT)cur_size, NULL);
|
|
*pfrag = 0;
|
|
bfd_section_size (stdoutput, now_seg) = size;
|
|
}
|
|
|
|
/* Switch back to current segment. */
|
|
subseg_set (current_seg, current_subseg);
|
|
|
|
#ifdef md_elf_section_change_hook
|
|
md_elf_section_change_hook ();
|
|
#endif
|
|
}
|
|
|
|
#endif /* TE_VMS */
|
|
|
|
/* Output COUNT bytes to a memory location. */
|
|
static char *vbyte_mem_ptr = NULL;
|
|
|
|
static void
|
|
output_vbyte_mem (int count, char *ptr, char *comment ATTRIBUTE_UNUSED)
|
|
{
|
|
int x;
|
|
if (vbyte_mem_ptr == NULL)
|
|
abort ();
|
|
|
|
if (count == 0)
|
|
return;
|
|
for (x = 0; x < count; x++)
|
|
*(vbyte_mem_ptr++) = ptr[x];
|
|
}
|
|
|
|
/* Count the number of bytes required for records. */
|
|
static int vbyte_count = 0;
|
|
static void
|
|
count_output (int count,
|
|
char *ptr ATTRIBUTE_UNUSED,
|
|
char *comment ATTRIBUTE_UNUSED)
|
|
{
|
|
vbyte_count += count;
|
|
}
|
|
|
|
static void
|
|
output_R1_format (vbyte_func f, unw_record_type rtype, int rlen)
|
|
{
|
|
int r = 0;
|
|
char byte;
|
|
if (rlen > 0x1f)
|
|
{
|
|
output_R3_format (f, rtype, rlen);
|
|
return;
|
|
}
|
|
|
|
if (rtype == body)
|
|
r = 1;
|
|
else if (rtype != prologue)
|
|
as_bad (_("record type is not valid"));
|
|
|
|
byte = UNW_R1 | (r << 5) | (rlen & 0x1f);
|
|
(*f) (1, &byte, NULL);
|
|
}
|
|
|
|
static void
|
|
output_R2_format (vbyte_func f, int mask, int grsave, unsigned long rlen)
|
|
{
|
|
char bytes[20];
|
|
int count = 2;
|
|
mask = (mask & 0x0f);
|
|
grsave = (grsave & 0x7f);
|
|
|
|
bytes[0] = (UNW_R2 | (mask >> 1));
|
|
bytes[1] = (((mask & 0x01) << 7) | grsave);
|
|
count += output_leb128 (bytes + 2, rlen, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_R3_format (vbyte_func f, unw_record_type rtype, unsigned long rlen)
|
|
{
|
|
int r = 0, count;
|
|
char bytes[20];
|
|
if (rlen <= 0x1f)
|
|
{
|
|
output_R1_format (f, rtype, rlen);
|
|
return;
|
|
}
|
|
|
|
if (rtype == body)
|
|
r = 1;
|
|
else if (rtype != prologue)
|
|
as_bad (_("record type is not valid"));
|
|
bytes[0] = (UNW_R3 | r);
|
|
count = output_leb128 (bytes + 1, rlen, 0);
|
|
(*f) (count + 1, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P1_format (vbyte_func f, int brmask)
|
|
{
|
|
char byte;
|
|
byte = UNW_P1 | (brmask & 0x1f);
|
|
(*f) (1, &byte, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P2_format (vbyte_func f, int brmask, int gr)
|
|
{
|
|
char bytes[2];
|
|
brmask = (brmask & 0x1f);
|
|
bytes[0] = UNW_P2 | (brmask >> 1);
|
|
bytes[1] = (((brmask & 1) << 7) | gr);
|
|
(*f) (2, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P3_format (vbyte_func f, unw_record_type rtype, int reg)
|
|
{
|
|
char bytes[2];
|
|
int r = 0;
|
|
reg = (reg & 0x7f);
|
|
switch (rtype)
|
|
{
|
|
case psp_gr:
|
|
r = 0;
|
|
break;
|
|
case rp_gr:
|
|
r = 1;
|
|
break;
|
|
case pfs_gr:
|
|
r = 2;
|
|
break;
|
|
case preds_gr:
|
|
r = 3;
|
|
break;
|
|
case unat_gr:
|
|
r = 4;
|
|
break;
|
|
case lc_gr:
|
|
r = 5;
|
|
break;
|
|
case rp_br:
|
|
r = 6;
|
|
break;
|
|
case rnat_gr:
|
|
r = 7;
|
|
break;
|
|
case bsp_gr:
|
|
r = 8;
|
|
break;
|
|
case bspstore_gr:
|
|
r = 9;
|
|
break;
|
|
case fpsr_gr:
|
|
r = 10;
|
|
break;
|
|
case priunat_gr:
|
|
r = 11;
|
|
break;
|
|
default:
|
|
as_bad (_("Invalid record type for P3 format."));
|
|
}
|
|
bytes[0] = (UNW_P3 | (r >> 1));
|
|
bytes[1] = (((r & 1) << 7) | reg);
|
|
(*f) (2, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P4_format (vbyte_func f, unsigned char *imask, unsigned long imask_size)
|
|
{
|
|
imask[0] = UNW_P4;
|
|
(*f) (imask_size, (char *) imask, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P5_format (vbyte_func f, int grmask, unsigned long frmask)
|
|
{
|
|
char bytes[4];
|
|
grmask = (grmask & 0x0f);
|
|
|
|
bytes[0] = UNW_P5;
|
|
bytes[1] = ((grmask << 4) | ((frmask & 0x000f0000) >> 16));
|
|
bytes[2] = ((frmask & 0x0000ff00) >> 8);
|
|
bytes[3] = (frmask & 0x000000ff);
|
|
(*f) (4, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P6_format (vbyte_func f, unw_record_type rtype, int rmask)
|
|
{
|
|
char byte;
|
|
int r = 0;
|
|
|
|
if (rtype == gr_mem)
|
|
r = 1;
|
|
else if (rtype != fr_mem)
|
|
as_bad (_("Invalid record type for format P6"));
|
|
byte = (UNW_P6 | (r << 4) | (rmask & 0x0f));
|
|
(*f) (1, &byte, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P7_format (vbyte_func f,
|
|
unw_record_type rtype,
|
|
unsigned long w1,
|
|
unsigned long w2)
|
|
{
|
|
char bytes[20];
|
|
int count = 1;
|
|
int r = 0;
|
|
count += output_leb128 (bytes + 1, w1, 0);
|
|
switch (rtype)
|
|
{
|
|
case mem_stack_f:
|
|
r = 0;
|
|
count += output_leb128 (bytes + count, w2 >> 4, 0);
|
|
break;
|
|
case mem_stack_v:
|
|
r = 1;
|
|
break;
|
|
case spill_base:
|
|
r = 2;
|
|
break;
|
|
case psp_sprel:
|
|
r = 3;
|
|
break;
|
|
case rp_when:
|
|
r = 4;
|
|
break;
|
|
case rp_psprel:
|
|
r = 5;
|
|
break;
|
|
case pfs_when:
|
|
r = 6;
|
|
break;
|
|
case pfs_psprel:
|
|
r = 7;
|
|
break;
|
|
case preds_when:
|
|
r = 8;
|
|
break;
|
|
case preds_psprel:
|
|
r = 9;
|
|
break;
|
|
case lc_when:
|
|
r = 10;
|
|
break;
|
|
case lc_psprel:
|
|
r = 11;
|
|
break;
|
|
case unat_when:
|
|
r = 12;
|
|
break;
|
|
case unat_psprel:
|
|
r = 13;
|
|
break;
|
|
case fpsr_when:
|
|
r = 14;
|
|
break;
|
|
case fpsr_psprel:
|
|
r = 15;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
bytes[0] = (UNW_P7 | r);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P8_format (vbyte_func f, unw_record_type rtype, unsigned long t)
|
|
{
|
|
char bytes[20];
|
|
int r = 0;
|
|
int count = 2;
|
|
bytes[0] = UNW_P8;
|
|
switch (rtype)
|
|
{
|
|
case rp_sprel:
|
|
r = 1;
|
|
break;
|
|
case pfs_sprel:
|
|
r = 2;
|
|
break;
|
|
case preds_sprel:
|
|
r = 3;
|
|
break;
|
|
case lc_sprel:
|
|
r = 4;
|
|
break;
|
|
case unat_sprel:
|
|
r = 5;
|
|
break;
|
|
case fpsr_sprel:
|
|
r = 6;
|
|
break;
|
|
case bsp_when:
|
|
r = 7;
|
|
break;
|
|
case bsp_psprel:
|
|
r = 8;
|
|
break;
|
|
case bsp_sprel:
|
|
r = 9;
|
|
break;
|
|
case bspstore_when:
|
|
r = 10;
|
|
break;
|
|
case bspstore_psprel:
|
|
r = 11;
|
|
break;
|
|
case bspstore_sprel:
|
|
r = 12;
|
|
break;
|
|
case rnat_when:
|
|
r = 13;
|
|
break;
|
|
case rnat_psprel:
|
|
r = 14;
|
|
break;
|
|
case rnat_sprel:
|
|
r = 15;
|
|
break;
|
|
case priunat_when_gr:
|
|
r = 16;
|
|
break;
|
|
case priunat_psprel:
|
|
r = 17;
|
|
break;
|
|
case priunat_sprel:
|
|
r = 18;
|
|
break;
|
|
case priunat_when_mem:
|
|
r = 19;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
bytes[1] = r;
|
|
count += output_leb128 (bytes + 2, t, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P9_format (vbyte_func f, int grmask, int gr)
|
|
{
|
|
char bytes[3];
|
|
bytes[0] = UNW_P9;
|
|
bytes[1] = (grmask & 0x0f);
|
|
bytes[2] = (gr & 0x7f);
|
|
(*f) (3, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_P10_format (vbyte_func f, int abi, int context)
|
|
{
|
|
char bytes[3];
|
|
bytes[0] = UNW_P10;
|
|
bytes[1] = (abi & 0xff);
|
|
bytes[2] = (context & 0xff);
|
|
(*f) (3, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_B1_format (vbyte_func f, unw_record_type rtype, unsigned long label)
|
|
{
|
|
char byte;
|
|
int r = 0;
|
|
if (label > 0x1f)
|
|
{
|
|
output_B4_format (f, rtype, label);
|
|
return;
|
|
}
|
|
if (rtype == copy_state)
|
|
r = 1;
|
|
else if (rtype != label_state)
|
|
as_bad (_("Invalid record type for format B1"));
|
|
|
|
byte = (UNW_B1 | (r << 5) | (label & 0x1f));
|
|
(*f) (1, &byte, NULL);
|
|
}
|
|
|
|
static void
|
|
output_B2_format (vbyte_func f, unsigned long ecount, unsigned long t)
|
|
{
|
|
char bytes[20];
|
|
int count = 1;
|
|
if (ecount > 0x1f)
|
|
{
|
|
output_B3_format (f, ecount, t);
|
|
return;
|
|
}
|
|
bytes[0] = (UNW_B2 | (ecount & 0x1f));
|
|
count += output_leb128 (bytes + 1, t, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_B3_format (vbyte_func f, unsigned long ecount, unsigned long t)
|
|
{
|
|
char bytes[20];
|
|
int count = 1;
|
|
if (ecount <= 0x1f)
|
|
{
|
|
output_B2_format (f, ecount, t);
|
|
return;
|
|
}
|
|
bytes[0] = UNW_B3;
|
|
count += output_leb128 (bytes + 1, t, 0);
|
|
count += output_leb128 (bytes + count, ecount, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_B4_format (vbyte_func f, unw_record_type rtype, unsigned long label)
|
|
{
|
|
char bytes[20];
|
|
int r = 0;
|
|
int count = 1;
|
|
if (label <= 0x1f)
|
|
{
|
|
output_B1_format (f, rtype, label);
|
|
return;
|
|
}
|
|
|
|
if (rtype == copy_state)
|
|
r = 1;
|
|
else if (rtype != label_state)
|
|
as_bad (_("Invalid record type for format B1"));
|
|
|
|
bytes[0] = (UNW_B4 | (r << 3));
|
|
count += output_leb128 (bytes + 1, label, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static char
|
|
format_ab_reg (int ab, int reg)
|
|
{
|
|
int ret;
|
|
ab = (ab & 3);
|
|
reg = (reg & 0x1f);
|
|
ret = (ab << 5) | reg;
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
output_X1_format (vbyte_func f,
|
|
unw_record_type rtype,
|
|
int ab,
|
|
int reg,
|
|
unsigned long t,
|
|
unsigned long w1)
|
|
{
|
|
char bytes[20];
|
|
int r = 0;
|
|
int count = 2;
|
|
bytes[0] = UNW_X1;
|
|
|
|
if (rtype == spill_sprel)
|
|
r = 1;
|
|
else if (rtype != spill_psprel)
|
|
as_bad (_("Invalid record type for format X1"));
|
|
bytes[1] = ((r << 7) | format_ab_reg (ab, reg));
|
|
count += output_leb128 (bytes + 2, t, 0);
|
|
count += output_leb128 (bytes + count, w1, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_X2_format (vbyte_func f,
|
|
int ab,
|
|
int reg,
|
|
int x,
|
|
int y,
|
|
int treg,
|
|
unsigned long t)
|
|
{
|
|
char bytes[20];
|
|
int count = 3;
|
|
bytes[0] = UNW_X2;
|
|
bytes[1] = (((x & 1) << 7) | format_ab_reg (ab, reg));
|
|
bytes[2] = (((y & 1) << 7) | (treg & 0x7f));
|
|
count += output_leb128 (bytes + 3, t, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_X3_format (vbyte_func f,
|
|
unw_record_type rtype,
|
|
int qp,
|
|
int ab,
|
|
int reg,
|
|
unsigned long t,
|
|
unsigned long w1)
|
|
{
|
|
char bytes[20];
|
|
int r = 0;
|
|
int count = 3;
|
|
bytes[0] = UNW_X3;
|
|
|
|
if (rtype == spill_sprel_p)
|
|
r = 1;
|
|
else if (rtype != spill_psprel_p)
|
|
as_bad (_("Invalid record type for format X3"));
|
|
bytes[1] = ((r << 7) | (qp & 0x3f));
|
|
bytes[2] = format_ab_reg (ab, reg);
|
|
count += output_leb128 (bytes + 3, t, 0);
|
|
count += output_leb128 (bytes + count, w1, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
static void
|
|
output_X4_format (vbyte_func f,
|
|
int qp,
|
|
int ab,
|
|
int reg,
|
|
int x,
|
|
int y,
|
|
int treg,
|
|
unsigned long t)
|
|
{
|
|
char bytes[20];
|
|
int count = 4;
|
|
bytes[0] = UNW_X4;
|
|
bytes[1] = (qp & 0x3f);
|
|
bytes[2] = (((x & 1) << 7) | format_ab_reg (ab, reg));
|
|
bytes[3] = (((y & 1) << 7) | (treg & 0x7f));
|
|
count += output_leb128 (bytes + 4, t, 0);
|
|
(*f) (count, bytes, NULL);
|
|
}
|
|
|
|
/* This function checks whether there are any outstanding .save-s and
|
|
discards them if so. */
|
|
|
|
static void
|
|
check_pending_save (void)
|
|
{
|
|
if (unwind.pending_saves)
|
|
{
|
|
unw_rec_list *cur, *prev;
|
|
|
|
as_warn (_("Previous .save incomplete"));
|
|
for (cur = unwind.list, prev = NULL; cur; )
|
|
if (&cur->r.record.p == unwind.pending_saves)
|
|
{
|
|
if (prev)
|
|
prev->next = cur->next;
|
|
else
|
|
unwind.list = cur->next;
|
|
if (cur == unwind.tail)
|
|
unwind.tail = prev;
|
|
if (cur == unwind.current_entry)
|
|
unwind.current_entry = cur->next;
|
|
/* Don't free the first discarded record, it's being used as
|
|
terminator for (currently) br_gr and gr_gr processing, and
|
|
also prevents leaving a dangling pointer to it in its
|
|
predecessor. */
|
|
cur->r.record.p.grmask = 0;
|
|
cur->r.record.p.brmask = 0;
|
|
cur->r.record.p.frmask = 0;
|
|
prev = cur->r.record.p.next;
|
|
cur->r.record.p.next = NULL;
|
|
cur = prev;
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
prev = cur;
|
|
cur = cur->next;
|
|
}
|
|
while (cur)
|
|
{
|
|
prev = cur;
|
|
cur = cur->r.record.p.next;
|
|
free (prev);
|
|
}
|
|
unwind.pending_saves = NULL;
|
|
}
|
|
}
|
|
|
|
/* This function allocates a record list structure, and initializes fields. */
|
|
|
|
static unw_rec_list *
|
|
alloc_record (unw_record_type t)
|
|
{
|
|
unw_rec_list *ptr;
|
|
ptr = XNEW (unw_rec_list);
|
|
memset (ptr, 0, sizeof (*ptr));
|
|
ptr->slot_number = SLOT_NUM_NOT_SET;
|
|
ptr->r.type = t;
|
|
return ptr;
|
|
}
|
|
|
|
/* Dummy unwind record used for calculating the length of the last prologue or
|
|
body region. */
|
|
|
|
static unw_rec_list *
|
|
output_endp (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (endp);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_prologue (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (prologue);
|
|
memset (&ptr->r.record.r.mask, 0, sizeof (ptr->r.record.r.mask));
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_prologue_gr (unsigned int saved_mask, unsigned int reg)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (prologue_gr);
|
|
memset (&ptr->r.record.r.mask, 0, sizeof (ptr->r.record.r.mask));
|
|
ptr->r.record.r.grmask = saved_mask;
|
|
ptr->r.record.r.grsave = reg;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_body (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (body);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_mem_stack_f (unsigned int size)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (mem_stack_f);
|
|
ptr->r.record.p.size = size;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_mem_stack_v (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (mem_stack_v);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_psp_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (psp_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_psp_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (psp_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rp_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rp_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rp_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rp_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rp_br (unsigned int br)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rp_br);
|
|
ptr->r.record.p.r.br = br;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rp_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rp_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rp_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rp_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_pfs_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (pfs_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_pfs_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (pfs_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_pfs_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (pfs_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_pfs_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (pfs_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_preds_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (preds_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_preds_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (preds_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_preds_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (preds_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_preds_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (preds_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_fr_mem (unsigned int mask)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (fr_mem);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
ptr->r.record.p.frmask = mask;
|
|
unwind.pending_saves = &ptr->r.record.p;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
mask &= ~(mask & (~mask + 1));
|
|
if (!mask)
|
|
return ptr;
|
|
cur = alloc_record (fr_mem);
|
|
cur->r.record.p.frmask = mask;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.frmask ^= mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_frgr_mem (unsigned int gr_mask, unsigned int fr_mask)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (frgr_mem);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
unwind.pending_saves = &cur->r.record.p;
|
|
cur->r.record.p.frmask = fr_mask;
|
|
while (fr_mask)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
fr_mask &= ~(fr_mask & (~fr_mask + 1));
|
|
if (!gr_mask && !fr_mask)
|
|
return ptr;
|
|
cur = alloc_record (frgr_mem);
|
|
cur->r.record.p.frmask = fr_mask;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.frmask ^= fr_mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
cur->r.record.p.grmask = gr_mask;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
gr_mask &= ~(gr_mask & (~gr_mask + 1));
|
|
if (!gr_mask)
|
|
return ptr;
|
|
cur = alloc_record (frgr_mem);
|
|
cur->r.record.p.grmask = gr_mask;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.grmask ^= gr_mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_gr_gr (unsigned int mask, unsigned int reg)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (gr_gr);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
ptr->r.record.p.grmask = mask;
|
|
ptr->r.record.p.r.gr = reg;
|
|
unwind.pending_saves = &ptr->r.record.p;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
mask &= ~(mask & (~mask + 1));
|
|
if (!mask)
|
|
return ptr;
|
|
cur = alloc_record (gr_gr);
|
|
cur->r.record.p.grmask = mask;
|
|
/* Indicate this record shouldn't be output. */
|
|
cur->r.record.p.r.gr = REG_NUM;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.grmask ^= mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_gr_mem (unsigned int mask)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (gr_mem);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
ptr->r.record.p.grmask = mask;
|
|
unwind.pending_saves = &ptr->r.record.p;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
mask &= ~(mask & (~mask + 1));
|
|
if (!mask)
|
|
return ptr;
|
|
cur = alloc_record (gr_mem);
|
|
cur->r.record.p.grmask = mask;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.grmask ^= mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_br_mem (unsigned int mask)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (br_mem);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
ptr->r.record.p.brmask = mask;
|
|
unwind.pending_saves = &ptr->r.record.p;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
mask &= ~(mask & (~mask + 1));
|
|
if (!mask)
|
|
return ptr;
|
|
cur = alloc_record (br_mem);
|
|
cur->r.record.p.brmask = mask;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.brmask ^= mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_br_gr (unsigned int mask, unsigned int reg)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (br_gr);
|
|
unw_rec_list *cur = ptr;
|
|
|
|
ptr->r.record.p.brmask = mask;
|
|
ptr->r.record.p.r.gr = reg;
|
|
unwind.pending_saves = &ptr->r.record.p;
|
|
for (;;)
|
|
{
|
|
unw_rec_list *prev = cur;
|
|
|
|
/* Clear least significant set bit. */
|
|
mask &= ~(mask & (~mask + 1));
|
|
if (!mask)
|
|
return ptr;
|
|
cur = alloc_record (br_gr);
|
|
cur->r.record.p.brmask = mask;
|
|
/* Indicate this record shouldn't be output. */
|
|
cur->r.record.p.r.gr = REG_NUM;
|
|
/* Retain only least significant bit. */
|
|
prev->r.record.p.brmask ^= mask;
|
|
prev->r.record.p.next = cur;
|
|
}
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_spill_base (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (spill_base);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_unat_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (unat_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_unat_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (unat_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_unat_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (unat_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_unat_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (unat_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_lc_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (lc_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_lc_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (lc_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_lc_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (lc_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_lc_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (lc_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_fpsr_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (fpsr_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_fpsr_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (fpsr_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_fpsr_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (fpsr_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_fpsr_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (fpsr_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_priunat_when_gr (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (priunat_when_gr);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_priunat_when_mem (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (priunat_when_mem);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_priunat_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (priunat_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_priunat_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (priunat_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_priunat_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (priunat_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bsp_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bsp_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bsp_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bsp_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bsp_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bsp_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bsp_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bsp_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bspstore_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bspstore_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bspstore_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bspstore_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bspstore_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bspstore_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_bspstore_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (bspstore_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rnat_when (void)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rnat_when);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rnat_gr (unsigned int gr)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rnat_gr);
|
|
ptr->r.record.p.r.gr = gr;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rnat_psprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rnat_psprel);
|
|
ptr->r.record.p.off.psp = ENCODED_PSP_OFFSET (offset);
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_rnat_sprel (unsigned int offset)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (rnat_sprel);
|
|
ptr->r.record.p.off.sp = offset / 4;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_unwabi (unsigned long abi, unsigned long context)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (unwabi);
|
|
ptr->r.record.p.abi = abi;
|
|
ptr->r.record.p.context = context;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_epilogue (unsigned long ecount)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (epilogue);
|
|
ptr->r.record.b.ecount = ecount;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_label_state (unsigned long label)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (label_state);
|
|
ptr->r.record.b.label = label;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_copy_state (unsigned long label)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (copy_state);
|
|
ptr->r.record.b.label = label;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_spill_psprel (unsigned int ab,
|
|
unsigned int reg,
|
|
unsigned int offset,
|
|
unsigned int predicate)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (predicate ? spill_psprel_p : spill_psprel);
|
|
ptr->r.record.x.ab = ab;
|
|
ptr->r.record.x.reg = reg;
|
|
ptr->r.record.x.where.pspoff = ENCODED_PSP_OFFSET (offset);
|
|
ptr->r.record.x.qp = predicate;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_spill_sprel (unsigned int ab,
|
|
unsigned int reg,
|
|
unsigned int offset,
|
|
unsigned int predicate)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (predicate ? spill_sprel_p : spill_sprel);
|
|
ptr->r.record.x.ab = ab;
|
|
ptr->r.record.x.reg = reg;
|
|
ptr->r.record.x.where.spoff = offset / 4;
|
|
ptr->r.record.x.qp = predicate;
|
|
return ptr;
|
|
}
|
|
|
|
static unw_rec_list *
|
|
output_spill_reg (unsigned int ab,
|
|
unsigned int reg,
|
|
unsigned int targ_reg,
|
|
unsigned int xy,
|
|
unsigned int predicate)
|
|
{
|
|
unw_rec_list *ptr = alloc_record (predicate ? spill_reg_p : spill_reg);
|
|
ptr->r.record.x.ab = ab;
|
|
ptr->r.record.x.reg = reg;
|
|
ptr->r.record.x.where.reg = targ_reg;
|
|
ptr->r.record.x.xy = xy;
|
|
ptr->r.record.x.qp = predicate;
|
|
return ptr;
|
|
}
|
|
|
|
/* Given a unw_rec_list process the correct format with the
|
|
specified function. */
|
|
|
|
static void
|
|
process_one_record (unw_rec_list *ptr, vbyte_func f)
|
|
{
|
|
unsigned int fr_mask, gr_mask;
|
|
|
|
switch (ptr->r.type)
|
|
{
|
|
/* This is a dummy record that takes up no space in the output. */
|
|
case endp:
|
|
break;
|
|
|
|
case gr_mem:
|
|
case fr_mem:
|
|
case br_mem:
|
|
case frgr_mem:
|
|
/* These are taken care of by prologue/prologue_gr. */
|
|
break;
|
|
|
|
case prologue_gr:
|
|
case prologue:
|
|
if (ptr->r.type == prologue_gr)
|
|
output_R2_format (f, ptr->r.record.r.grmask,
|
|
ptr->r.record.r.grsave, ptr->r.record.r.rlen);
|
|
else
|
|
output_R1_format (f, ptr->r.type, ptr->r.record.r.rlen);
|
|
|
|
/* Output descriptor(s) for union of register spills (if any). */
|
|
gr_mask = ptr->r.record.r.mask.gr_mem;
|
|
fr_mask = ptr->r.record.r.mask.fr_mem;
|
|
if (fr_mask)
|
|
{
|
|
if ((fr_mask & ~0xfUL) == 0)
|
|
output_P6_format (f, fr_mem, fr_mask);
|
|
else
|
|
{
|
|
output_P5_format (f, gr_mask, fr_mask);
|
|
gr_mask = 0;
|
|
}
|
|
}
|
|
if (gr_mask)
|
|
output_P6_format (f, gr_mem, gr_mask);
|
|
if (ptr->r.record.r.mask.br_mem)
|
|
output_P1_format (f, ptr->r.record.r.mask.br_mem);
|
|
|
|
/* output imask descriptor if necessary: */
|
|
if (ptr->r.record.r.mask.i)
|
|
output_P4_format (f, ptr->r.record.r.mask.i,
|
|
ptr->r.record.r.imask_size);
|
|
break;
|
|
|
|
case body:
|
|
output_R1_format (f, ptr->r.type, ptr->r.record.r.rlen);
|
|
break;
|
|
case mem_stack_f:
|
|
case mem_stack_v:
|
|
output_P7_format (f, ptr->r.type, ptr->r.record.p.t,
|
|
ptr->r.record.p.size);
|
|
break;
|
|
case psp_gr:
|
|
case rp_gr:
|
|
case pfs_gr:
|
|
case preds_gr:
|
|
case unat_gr:
|
|
case lc_gr:
|
|
case fpsr_gr:
|
|
case priunat_gr:
|
|
case bsp_gr:
|
|
case bspstore_gr:
|
|
case rnat_gr:
|
|
output_P3_format (f, ptr->r.type, ptr->r.record.p.r.gr);
|
|
break;
|
|
case rp_br:
|
|
output_P3_format (f, rp_br, ptr->r.record.p.r.br);
|
|
break;
|
|
case psp_sprel:
|
|
output_P7_format (f, psp_sprel, ptr->r.record.p.off.sp, 0);
|
|
break;
|
|
case rp_when:
|
|
case pfs_when:
|
|
case preds_when:
|
|
case unat_when:
|
|
case lc_when:
|
|
case fpsr_when:
|
|
output_P7_format (f, ptr->r.type, ptr->r.record.p.t, 0);
|
|
break;
|
|
case rp_psprel:
|
|
case pfs_psprel:
|
|
case preds_psprel:
|
|
case unat_psprel:
|
|
case lc_psprel:
|
|
case fpsr_psprel:
|
|
case spill_base:
|
|
output_P7_format (f, ptr->r.type, ptr->r.record.p.off.psp, 0);
|
|
break;
|
|
case rp_sprel:
|
|
case pfs_sprel:
|
|
case preds_sprel:
|
|
case unat_sprel:
|
|
case lc_sprel:
|
|
case fpsr_sprel:
|
|
case priunat_sprel:
|
|
case bsp_sprel:
|
|
case bspstore_sprel:
|
|
case rnat_sprel:
|
|
output_P8_format (f, ptr->r.type, ptr->r.record.p.off.sp);
|
|
break;
|
|
case gr_gr:
|
|
if (ptr->r.record.p.r.gr < REG_NUM)
|
|
{
|
|
const unw_rec_list *cur = ptr;
|
|
|
|
gr_mask = cur->r.record.p.grmask;
|
|
while ((cur = cur->r.record.p.next) != NULL)
|
|
gr_mask |= cur->r.record.p.grmask;
|
|
output_P9_format (f, gr_mask, ptr->r.record.p.r.gr);
|
|
}
|
|
break;
|
|
case br_gr:
|
|
if (ptr->r.record.p.r.gr < REG_NUM)
|
|
{
|
|
const unw_rec_list *cur = ptr;
|
|
|
|
gr_mask = cur->r.record.p.brmask;
|
|
while ((cur = cur->r.record.p.next) != NULL)
|
|
gr_mask |= cur->r.record.p.brmask;
|
|
output_P2_format (f, gr_mask, ptr->r.record.p.r.gr);
|
|
}
|
|
break;
|
|
case spill_mask:
|
|
as_bad (_("spill_mask record unimplemented."));
|
|
break;
|
|
case priunat_when_gr:
|
|
case priunat_when_mem:
|
|
case bsp_when:
|
|
case bspstore_when:
|
|
case rnat_when:
|
|
output_P8_format (f, ptr->r.type, ptr->r.record.p.t);
|
|
break;
|
|
case priunat_psprel:
|
|
case bsp_psprel:
|
|
case bspstore_psprel:
|
|
case rnat_psprel:
|
|
output_P8_format (f, ptr->r.type, ptr->r.record.p.off.psp);
|
|
break;
|
|
case unwabi:
|
|
output_P10_format (f, ptr->r.record.p.abi, ptr->r.record.p.context);
|
|
break;
|
|
case epilogue:
|
|
output_B3_format (f, ptr->r.record.b.ecount, ptr->r.record.b.t);
|
|
break;
|
|
case label_state:
|
|
case copy_state:
|
|
output_B4_format (f, ptr->r.type, ptr->r.record.b.label);
|
|
break;
|
|
case spill_psprel:
|
|
output_X1_format (f, ptr->r.type, ptr->r.record.x.ab,
|
|
ptr->r.record.x.reg, ptr->r.record.x.t,
|
|
ptr->r.record.x.where.pspoff);
|
|
break;
|
|
case spill_sprel:
|
|
output_X1_format (f, ptr->r.type, ptr->r.record.x.ab,
|
|
ptr->r.record.x.reg, ptr->r.record.x.t,
|
|
ptr->r.record.x.where.spoff);
|
|
break;
|
|
case spill_reg:
|
|
output_X2_format (f, ptr->r.record.x.ab, ptr->r.record.x.reg,
|
|
ptr->r.record.x.xy >> 1, ptr->r.record.x.xy,
|
|
ptr->r.record.x.where.reg, ptr->r.record.x.t);
|
|
break;
|
|
case spill_psprel_p:
|
|
output_X3_format (f, ptr->r.type, ptr->r.record.x.qp,
|
|
ptr->r.record.x.ab, ptr->r.record.x.reg,
|
|
ptr->r.record.x.t, ptr->r.record.x.where.pspoff);
|
|
break;
|
|
case spill_sprel_p:
|
|
output_X3_format (f, ptr->r.type, ptr->r.record.x.qp,
|
|
ptr->r.record.x.ab, ptr->r.record.x.reg,
|
|
ptr->r.record.x.t, ptr->r.record.x.where.spoff);
|
|
break;
|
|
case spill_reg_p:
|
|
output_X4_format (f, ptr->r.record.x.qp, ptr->r.record.x.ab,
|
|
ptr->r.record.x.reg, ptr->r.record.x.xy >> 1,
|
|
ptr->r.record.x.xy, ptr->r.record.x.where.reg,
|
|
ptr->r.record.x.t);
|
|
break;
|
|
default:
|
|
as_bad (_("record_type_not_valid"));
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Given a unw_rec_list list, process all the records with
|
|
the specified function. */
|
|
static void
|
|
process_unw_records (unw_rec_list *list, vbyte_func f)
|
|
{
|
|
unw_rec_list *ptr;
|
|
for (ptr = list; ptr; ptr = ptr->next)
|
|
process_one_record (ptr, f);
|
|
}
|
|
|
|
/* Determine the size of a record list in bytes. */
|
|
static int
|
|
calc_record_size (unw_rec_list *list)
|
|
{
|
|
vbyte_count = 0;
|
|
process_unw_records (list, count_output);
|
|
return vbyte_count;
|
|
}
|
|
|
|
/* Return the number of bits set in the input value.
|
|
Perhaps this has a better place... */
|
|
#if __GNUC__ > 3 || (__GNUC__ == 3 && __GNUC_MINOR__ >= 4)
|
|
# define popcount __builtin_popcount
|
|
#else
|
|
static int
|
|
popcount (unsigned x)
|
|
{
|
|
static const unsigned char popcnt[16] =
|
|
{
|
|
0, 1, 1, 2,
|
|
1, 2, 2, 3,
|
|
1, 2, 2, 3,
|
|
2, 3, 3, 4
|
|
};
|
|
|
|
if (x < NELEMS (popcnt))
|
|
return popcnt[x];
|
|
return popcnt[x % NELEMS (popcnt)] + popcount (x / NELEMS (popcnt));
|
|
}
|
|
#endif
|
|
|
|
/* Update IMASK bitmask to reflect the fact that one or more registers
|
|
of type TYPE are saved starting at instruction with index T. If N
|
|
bits are set in REGMASK, it is assumed that instructions T through
|
|
T+N-1 save these registers.
|
|
|
|
TYPE values:
|
|
0: no save
|
|
1: instruction saves next fp reg
|
|
2: instruction saves next general reg
|
|
3: instruction saves next branch reg */
|
|
static void
|
|
set_imask (unw_rec_list *region,
|
|
unsigned long regmask,
|
|
unsigned long t,
|
|
unsigned int type)
|
|
{
|
|
unsigned char *imask;
|
|
unsigned long imask_size;
|
|
unsigned int i;
|
|
int pos;
|
|
|
|
imask = region->r.record.r.mask.i;
|
|
imask_size = region->r.record.r.imask_size;
|
|
if (!imask)
|
|
{
|
|
imask_size = (region->r.record.r.rlen * 2 + 7) / 8 + 1;
|
|
imask = XCNEWVEC (unsigned char, imask_size);
|
|
|
|
region->r.record.r.imask_size = imask_size;
|
|
region->r.record.r.mask.i = imask;
|
|
}
|
|
|
|
i = (t / 4) + 1;
|
|
pos = 2 * (3 - t % 4);
|
|
while (regmask)
|
|
{
|
|
if (i >= imask_size)
|
|
{
|
|
as_bad (_("Ignoring attempt to spill beyond end of region"));
|
|
return;
|
|
}
|
|
|
|
imask[i] |= (type & 0x3) << pos;
|
|
|
|
regmask &= (regmask - 1);
|
|
pos -= 2;
|
|
if (pos < 0)
|
|
{
|
|
pos = 0;
|
|
++i;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Return the number of instruction slots from FIRST_ADDR to SLOT_ADDR.
|
|
SLOT_FRAG is the frag containing SLOT_ADDR, and FIRST_FRAG is the frag
|
|
containing FIRST_ADDR. If BEFORE_RELAX, then we use worst-case estimates
|
|
for frag sizes. */
|
|
|
|
static unsigned long
|
|
slot_index (unsigned long slot_addr,
|
|
fragS *slot_frag,
|
|
unsigned long first_addr,
|
|
fragS *first_frag,
|
|
int before_relax)
|
|
{
|
|
unsigned long s_index = 0;
|
|
|
|
/* First time we are called, the initial address and frag are invalid. */
|
|
if (first_addr == 0)
|
|
return 0;
|
|
|
|
/* If the two addresses are in different frags, then we need to add in
|
|
the remaining size of this frag, and then the entire size of intermediate
|
|
frags. */
|
|
while (slot_frag != first_frag)
|
|
{
|
|
unsigned long start_addr = (unsigned long) &first_frag->fr_literal;
|
|
|
|
if (! before_relax)
|
|
{
|
|
/* We can get the final addresses only during and after
|
|
relaxation. */
|
|
if (first_frag->fr_next && first_frag->fr_next->fr_address)
|
|
s_index += 3 * ((first_frag->fr_next->fr_address
|
|
- first_frag->fr_address
|
|
- first_frag->fr_fix) >> 4);
|
|
}
|
|
else
|
|
/* We don't know what the final addresses will be. We try our
|
|
best to estimate. */
|
|
switch (first_frag->fr_type)
|
|
{
|
|
default:
|
|
break;
|
|
|
|
case rs_space:
|
|
as_fatal (_("Only constant space allocation is supported"));
|
|
break;
|
|
|
|
case rs_align:
|
|
case rs_align_code:
|
|
case rs_align_test:
|
|
/* Take alignment into account. Assume the worst case
|
|
before relaxation. */
|
|
s_index += 3 * ((1 << first_frag->fr_offset) >> 4);
|
|
break;
|
|
|
|
case rs_org:
|
|
if (first_frag->fr_symbol)
|
|
{
|
|
as_fatal (_("Only constant offsets are supported"));
|
|
break;
|
|
}
|
|
/* Fall through. */
|
|
case rs_fill:
|
|
s_index += 3 * (first_frag->fr_offset >> 4);
|
|
break;
|
|
}
|
|
|
|
/* Add in the full size of the frag converted to instruction slots. */
|
|
s_index += 3 * (first_frag->fr_fix >> 4);
|
|
/* Subtract away the initial part before first_addr. */
|
|
s_index -= (3 * ((first_addr >> 4) - (start_addr >> 4))
|
|
+ ((first_addr & 0x3) - (start_addr & 0x3)));
|
|
|
|
/* Move to the beginning of the next frag. */
|
|
first_frag = first_frag->fr_next;
|
|
first_addr = (unsigned long) &first_frag->fr_literal;
|
|
|
|
/* This can happen if there is section switching in the middle of a
|
|
function, causing the frag chain for the function to be broken.
|
|
It is too difficult to recover safely from this problem, so we just
|
|
exit with an error. */
|
|
if (first_frag == NULL)
|
|
as_fatal (_("Section switching in code is not supported."));
|
|
}
|
|
|
|
/* Add in the used part of the last frag. */
|
|
s_index += (3 * ((slot_addr >> 4) - (first_addr >> 4))
|
|
+ ((slot_addr & 0x3) - (first_addr & 0x3)));
|
|
return s_index;
|
|
}
|
|
|
|
/* Optimize unwind record directives. */
|
|
|
|
static unw_rec_list *
|
|
optimize_unw_records (unw_rec_list *list)
|
|
{
|
|
if (!list)
|
|
return NULL;
|
|
|
|
/* If the only unwind record is ".prologue" or ".prologue" followed
|
|
by ".body", then we can optimize the unwind directives away. */
|
|
if (list->r.type == prologue
|
|
&& (list->next->r.type == endp
|
|
|| (list->next->r.type == body && list->next->next->r.type == endp)))
|
|
return NULL;
|
|
|
|
return list;
|
|
}
|
|
|
|
/* Given a complete record list, process any records which have
|
|
unresolved fields, (ie length counts for a prologue). After
|
|
this has been run, all necessary information should be available
|
|
within each record to generate an image. */
|
|
|
|
static void
|
|
fixup_unw_records (unw_rec_list *list, int before_relax)
|
|
{
|
|
unw_rec_list *ptr, *region = 0;
|
|
unsigned long first_addr = 0, rlen = 0, t;
|
|
fragS *first_frag = 0;
|
|
|
|
for (ptr = list; ptr; ptr = ptr->next)
|
|
{
|
|
if (ptr->slot_number == SLOT_NUM_NOT_SET)
|
|
as_bad (_("Insn slot not set in unwind record."));
|
|
t = slot_index (ptr->slot_number, ptr->slot_frag,
|
|
first_addr, first_frag, before_relax);
|
|
switch (ptr->r.type)
|
|
{
|
|
case prologue:
|
|
case prologue_gr:
|
|
case body:
|
|
{
|
|
unw_rec_list *last;
|
|
int size;
|
|
unsigned long last_addr = 0;
|
|
fragS *last_frag = NULL;
|
|
|
|
first_addr = ptr->slot_number;
|
|
first_frag = ptr->slot_frag;
|
|
/* Find either the next body/prologue start, or the end of
|
|
the function, and determine the size of the region. */
|
|
for (last = ptr->next; last != NULL; last = last->next)
|
|
if (last->r.type == prologue || last->r.type == prologue_gr
|
|
|| last->r.type == body || last->r.type == endp)
|
|
{
|
|
last_addr = last->slot_number;
|
|
last_frag = last->slot_frag;
|
|
break;
|
|
}
|
|
size = slot_index (last_addr, last_frag, first_addr, first_frag,
|
|
before_relax);
|
|
rlen = ptr->r.record.r.rlen = size;
|
|
if (ptr->r.type == body)
|
|
/* End of region. */
|
|
region = 0;
|
|
else
|
|
region = ptr;
|
|
break;
|
|
}
|
|
case epilogue:
|
|
if (t < rlen)
|
|
ptr->r.record.b.t = rlen - 1 - t;
|
|
else
|
|
/* This happens when a memory-stack-less procedure uses a
|
|
".restore sp" directive at the end of a region to pop
|
|
the frame state. */
|
|
ptr->r.record.b.t = 0;
|
|
break;
|
|
|
|
case mem_stack_f:
|
|
case mem_stack_v:
|
|
case rp_when:
|
|
case pfs_when:
|
|
case preds_when:
|
|
case unat_when:
|
|
case lc_when:
|
|
case fpsr_when:
|
|
case priunat_when_gr:
|
|
case priunat_when_mem:
|
|
case bsp_when:
|
|
case bspstore_when:
|
|
case rnat_when:
|
|
ptr->r.record.p.t = t;
|
|
break;
|
|
|
|
case spill_reg:
|
|
case spill_sprel:
|
|
case spill_psprel:
|
|
case spill_reg_p:
|
|
case spill_sprel_p:
|
|
case spill_psprel_p:
|
|
ptr->r.record.x.t = t;
|
|
break;
|
|
|
|
case frgr_mem:
|
|
if (!region)
|
|
{
|
|
as_bad (_("frgr_mem record before region record!"));
|
|
return;
|
|
}
|
|
region->r.record.r.mask.fr_mem |= ptr->r.record.p.frmask;
|
|
region->r.record.r.mask.gr_mem |= ptr->r.record.p.grmask;
|
|
set_imask (region, ptr->r.record.p.frmask, t, 1);
|
|
set_imask (region, ptr->r.record.p.grmask, t, 2);
|
|
break;
|
|
case fr_mem:
|
|
if (!region)
|
|
{
|
|
as_bad (_("fr_mem record before region record!"));
|
|
return;
|
|
}
|
|
region->r.record.r.mask.fr_mem |= ptr->r.record.p.frmask;
|
|
set_imask (region, ptr->r.record.p.frmask, t, 1);
|
|
break;
|
|
case gr_mem:
|
|
if (!region)
|
|
{
|
|
as_bad (_("gr_mem record before region record!"));
|
|
return;
|
|
}
|
|
region->r.record.r.mask.gr_mem |= ptr->r.record.p.grmask;
|
|
set_imask (region, ptr->r.record.p.grmask, t, 2);
|
|
break;
|
|
case br_mem:
|
|
if (!region)
|
|
{
|
|
as_bad (_("br_mem record before region record!"));
|
|
return;
|
|
}
|
|
region->r.record.r.mask.br_mem |= ptr->r.record.p.brmask;
|
|
set_imask (region, ptr->r.record.p.brmask, t, 3);
|
|
break;
|
|
|
|
case gr_gr:
|
|
if (!region)
|
|
{
|
|
as_bad (_("gr_gr record before region record!"));
|
|
return;
|
|
}
|
|
set_imask (region, ptr->r.record.p.grmask, t, 2);
|
|
break;
|
|
case br_gr:
|
|
if (!region)
|
|
{
|
|
as_bad (_("br_gr record before region record!"));
|
|
return;
|
|
}
|
|
set_imask (region, ptr->r.record.p.brmask, t, 3);
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Estimate the size of a frag before relaxing. We only have one type of frag
|
|
to handle here, which is the unwind info frag. */
|
|
|
|
int
|
|
ia64_estimate_size_before_relax (fragS *frag,
|
|
asection *segtype ATTRIBUTE_UNUSED)
|
|
{
|
|
unw_rec_list *list;
|
|
int len, size, pad;
|
|
|
|
/* ??? This code is identical to the first part of ia64_convert_frag. */
|
|
list = (unw_rec_list *) frag->fr_opcode;
|
|
fixup_unw_records (list, 0);
|
|
|
|
len = calc_record_size (list);
|
|
/* pad to pointer-size boundary. */
|
|
pad = len % md.pointer_size;
|
|
if (pad != 0)
|
|
len += md.pointer_size - pad;
|
|
/* Add 8 for the header. */
|
|
size = len + 8;
|
|
/* Add a pointer for the personality offset. */
|
|
if (frag->fr_offset)
|
|
size += md.pointer_size;
|
|
|
|
/* fr_var carries the max_chars that we created the fragment with.
|
|
We must, of course, have allocated enough memory earlier. */
|
|
gas_assert (frag->fr_var >= size);
|
|
|
|
return frag->fr_fix + size;
|
|
}
|
|
|
|
/* This function converts a rs_machine_dependent variant frag into a
|
|
normal fill frag with the unwind image from the record list. */
|
|
void
|
|
ia64_convert_frag (fragS *frag)
|
|
{
|
|
unw_rec_list *list;
|
|
int len, size, pad;
|
|
valueT flag_value;
|
|
|
|
/* ??? This code is identical to ia64_estimate_size_before_relax. */
|
|
list = (unw_rec_list *) frag->fr_opcode;
|
|
fixup_unw_records (list, 0);
|
|
|
|
len = calc_record_size (list);
|
|
/* pad to pointer-size boundary. */
|
|
pad = len % md.pointer_size;
|
|
if (pad != 0)
|
|
len += md.pointer_size - pad;
|
|
/* Add 8 for the header. */
|
|
size = len + 8;
|
|
/* Add a pointer for the personality offset. */
|
|
if (frag->fr_offset)
|
|
size += md.pointer_size;
|
|
|
|
/* fr_var carries the max_chars that we created the fragment with.
|
|
We must, of course, have allocated enough memory earlier. */
|
|
gas_assert (frag->fr_var >= size);
|
|
|
|
/* Initialize the header area. fr_offset is initialized with
|
|
unwind.personality_routine. */
|
|
if (frag->fr_offset)
|
|
{
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
flag_value = (bfd_vma) 3 << 32;
|
|
else
|
|
/* 32-bit unwind info block. */
|
|
flag_value = (bfd_vma) 0x1003 << 32;
|
|
}
|
|
else
|
|
flag_value = 0;
|
|
|
|
md_number_to_chars (frag->fr_literal,
|
|
(((bfd_vma) 1 << 48) /* Version. */
|
|
| flag_value /* U & E handler flags. */
|
|
| (len / md.pointer_size)), /* Length. */
|
|
8);
|
|
|
|
/* Skip the header. */
|
|
vbyte_mem_ptr = frag->fr_literal + 8;
|
|
process_unw_records (list, output_vbyte_mem);
|
|
|
|
/* Fill the padding bytes with zeros. */
|
|
if (pad != 0)
|
|
md_number_to_chars (frag->fr_literal + len + 8 - md.pointer_size + pad, 0,
|
|
md.pointer_size - pad);
|
|
/* Fill the unwind personality with zeros. */
|
|
if (frag->fr_offset)
|
|
md_number_to_chars (frag->fr_literal + size - md.pointer_size, 0,
|
|
md.pointer_size);
|
|
|
|
frag->fr_fix += size;
|
|
frag->fr_type = rs_fill;
|
|
frag->fr_var = 0;
|
|
frag->fr_offset = 0;
|
|
}
|
|
|
|
static int
|
|
parse_predicate_and_operand (expressionS *e, unsigned *qp, const char *po)
|
|
{
|
|
int sep = parse_operand_and_eval (e, ',');
|
|
|
|
*qp = e->X_add_number - REG_P;
|
|
if (e->X_op != O_register || *qp > 63)
|
|
{
|
|
as_bad (_("First operand to .%s must be a predicate"), po);
|
|
*qp = 0;
|
|
}
|
|
else if (*qp == 0)
|
|
as_warn (_("Pointless use of p0 as first operand to .%s"), po);
|
|
if (sep == ',')
|
|
sep = parse_operand_and_eval (e, ',');
|
|
else
|
|
e->X_op = O_absent;
|
|
return sep;
|
|
}
|
|
|
|
static void
|
|
convert_expr_to_ab_reg (const expressionS *e,
|
|
unsigned int *ab,
|
|
unsigned int *regp,
|
|
const char *po,
|
|
int n)
|
|
{
|
|
unsigned int reg = e->X_add_number;
|
|
|
|
*ab = *regp = 0; /* Anything valid is good here. */
|
|
|
|
if (e->X_op != O_register)
|
|
reg = REG_GR; /* Anything invalid is good here. */
|
|
|
|
if (reg >= (REG_GR + 4) && reg <= (REG_GR + 7))
|
|
{
|
|
*ab = 0;
|
|
*regp = reg - REG_GR;
|
|
}
|
|
else if ((reg >= (REG_FR + 2) && reg <= (REG_FR + 5))
|
|
|| (reg >= (REG_FR + 16) && reg <= (REG_FR + 31)))
|
|
{
|
|
*ab = 1;
|
|
*regp = reg - REG_FR;
|
|
}
|
|
else if (reg >= (REG_BR + 1) && reg <= (REG_BR + 5))
|
|
{
|
|
*ab = 2;
|
|
*regp = reg - REG_BR;
|
|
}
|
|
else
|
|
{
|
|
*ab = 3;
|
|
switch (reg)
|
|
{
|
|
case REG_PR: *regp = 0; break;
|
|
case REG_PSP: *regp = 1; break;
|
|
case REG_PRIUNAT: *regp = 2; break;
|
|
case REG_BR + 0: *regp = 3; break;
|
|
case REG_AR + AR_BSP: *regp = 4; break;
|
|
case REG_AR + AR_BSPSTORE: *regp = 5; break;
|
|
case REG_AR + AR_RNAT: *regp = 6; break;
|
|
case REG_AR + AR_UNAT: *regp = 7; break;
|
|
case REG_AR + AR_FPSR: *regp = 8; break;
|
|
case REG_AR + AR_PFS: *regp = 9; break;
|
|
case REG_AR + AR_LC: *regp = 10; break;
|
|
|
|
default:
|
|
as_bad (_("Operand %d to .%s must be a preserved register"), n, po);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
convert_expr_to_xy_reg (const expressionS *e,
|
|
unsigned int *xy,
|
|
unsigned int *regp,
|
|
const char *po,
|
|
int n)
|
|
{
|
|
unsigned int reg = e->X_add_number;
|
|
|
|
*xy = *regp = 0; /* Anything valid is good here. */
|
|
|
|
if (e->X_op != O_register)
|
|
reg = REG_GR; /* Anything invalid is good here. */
|
|
|
|
if (reg >= (REG_GR + 1) && reg <= (REG_GR + 127))
|
|
{
|
|
*xy = 0;
|
|
*regp = reg - REG_GR;
|
|
}
|
|
else if (reg >= (REG_FR + 2) && reg <= (REG_FR + 127))
|
|
{
|
|
*xy = 1;
|
|
*regp = reg - REG_FR;
|
|
}
|
|
else if (reg >= REG_BR && reg <= (REG_BR + 7))
|
|
{
|
|
*xy = 2;
|
|
*regp = reg - REG_BR;
|
|
}
|
|
else
|
|
as_bad (_("Operand %d to .%s must be a writable register"), n, po);
|
|
}
|
|
|
|
static void
|
|
dot_align (int arg)
|
|
{
|
|
/* The current frag is an alignment frag. */
|
|
align_frag = frag_now;
|
|
s_align_bytes (arg);
|
|
}
|
|
|
|
static void
|
|
dot_radix (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
char *radix;
|
|
int ch;
|
|
|
|
SKIP_WHITESPACE ();
|
|
|
|
if (is_it_end_of_statement ())
|
|
return;
|
|
ch = get_symbol_name (&radix);
|
|
ia64_canonicalize_symbol_name (radix);
|
|
if (strcasecmp (radix, "C"))
|
|
as_bad (_("Radix `%s' unsupported or invalid"), radix);
|
|
(void) restore_line_pointer (ch);
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* Helper function for .loc directives. If the assembler is not generating
|
|
line number info, then we need to remember which instructions have a .loc
|
|
directive, and only call dwarf2_gen_line_info for those instructions. */
|
|
|
|
static void
|
|
dot_loc (int x)
|
|
{
|
|
CURR_SLOT.loc_directive_seen = 1;
|
|
dwarf2_directive_loc (x);
|
|
}
|
|
|
|
/* .sbss, .bss etc. are macros that expand into ".section SECNAME". */
|
|
static void
|
|
dot_special_section (int which)
|
|
{
|
|
set_section ((char *) special_section_name[which]);
|
|
}
|
|
|
|
/* Return -1 for warning and 0 for error. */
|
|
|
|
static int
|
|
unwind_diagnostic (const char * region, const char *directive)
|
|
{
|
|
if (md.unwind_check == unwind_check_warning)
|
|
{
|
|
as_warn (_(".%s outside of %s"), directive, region);
|
|
return -1;
|
|
}
|
|
else
|
|
{
|
|
as_bad (_(".%s outside of %s"), directive, region);
|
|
ignore_rest_of_line ();
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* Return 1 if a directive is in a procedure, -1 if a directive isn't in
|
|
a procedure but the unwind directive check is set to warning, 0 if
|
|
a directive isn't in a procedure and the unwind directive check is set
|
|
to error. */
|
|
|
|
static int
|
|
in_procedure (const char *directive)
|
|
{
|
|
if (unwind.proc_pending.sym
|
|
&& (!unwind.saved_text_seg || strcmp (directive, "endp") == 0))
|
|
return 1;
|
|
return unwind_diagnostic ("procedure", directive);
|
|
}
|
|
|
|
/* Return 1 if a directive is in a prologue, -1 if a directive isn't in
|
|
a prologue but the unwind directive check is set to warning, 0 if
|
|
a directive isn't in a prologue and the unwind directive check is set
|
|
to error. */
|
|
|
|
static int
|
|
in_prologue (const char *directive)
|
|
{
|
|
int in = in_procedure (directive);
|
|
|
|
if (in > 0 && !unwind.prologue)
|
|
in = unwind_diagnostic ("prologue", directive);
|
|
check_pending_save ();
|
|
return in;
|
|
}
|
|
|
|
/* Return 1 if a directive is in a body, -1 if a directive isn't in
|
|
a body but the unwind directive check is set to warning, 0 if
|
|
a directive isn't in a body and the unwind directive check is set
|
|
to error. */
|
|
|
|
static int
|
|
in_body (const char *directive)
|
|
{
|
|
int in = in_procedure (directive);
|
|
|
|
if (in > 0 && !unwind.body)
|
|
in = unwind_diagnostic ("body region", directive);
|
|
return in;
|
|
}
|
|
|
|
static void
|
|
add_unwind_entry (unw_rec_list *ptr, int sep)
|
|
{
|
|
if (ptr)
|
|
{
|
|
if (unwind.tail)
|
|
unwind.tail->next = ptr;
|
|
else
|
|
unwind.list = ptr;
|
|
unwind.tail = ptr;
|
|
|
|
/* The current entry can in fact be a chain of unwind entries. */
|
|
if (unwind.current_entry == NULL)
|
|
unwind.current_entry = ptr;
|
|
}
|
|
|
|
/* The current entry can in fact be a chain of unwind entries. */
|
|
if (unwind.current_entry == NULL)
|
|
unwind.current_entry = ptr;
|
|
|
|
if (sep == ',')
|
|
{
|
|
char *name;
|
|
/* Parse a tag permitted for the current directive. */
|
|
int ch;
|
|
|
|
SKIP_WHITESPACE ();
|
|
ch = get_symbol_name (&name);
|
|
/* FIXME: For now, just issue a warning that this isn't implemented. */
|
|
{
|
|
static int warned;
|
|
|
|
if (!warned)
|
|
{
|
|
warned = 1;
|
|
as_warn (_("Tags on unwind pseudo-ops aren't supported, yet"));
|
|
}
|
|
}
|
|
(void) restore_line_pointer (ch);
|
|
}
|
|
if (sep != NOT_A_CHAR)
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_fframe (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
int sep;
|
|
|
|
if (!in_prologue ("fframe"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
|
|
if (e.X_op != O_constant)
|
|
{
|
|
as_bad (_("First operand to .fframe must be a constant"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_mem_stack_f (e.X_add_number), sep);
|
|
}
|
|
|
|
static void
|
|
dot_vframe (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
unsigned reg;
|
|
int sep;
|
|
|
|
if (!in_prologue ("vframe"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
reg = e.X_add_number - REG_GR;
|
|
if (e.X_op != O_register || reg > 127)
|
|
{
|
|
as_bad (_("First operand to .vframe must be a general register"));
|
|
reg = 0;
|
|
}
|
|
add_unwind_entry (output_mem_stack_v (), sep);
|
|
if (! (unwind.prologue_mask & 2))
|
|
add_unwind_entry (output_psp_gr (reg), NOT_A_CHAR);
|
|
else if (reg != unwind.prologue_gr
|
|
+ (unsigned) popcount (unwind.prologue_mask & -(2 << 1)))
|
|
as_warn (_("Operand of .vframe contradicts .prologue"));
|
|
}
|
|
|
|
static void
|
|
dot_vframesp (int psp)
|
|
{
|
|
expressionS e;
|
|
int sep;
|
|
|
|
if (psp)
|
|
as_warn (_(".vframepsp is meaningless, assuming .vframesp was meant"));
|
|
|
|
if (!in_prologue ("vframesp"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
if (e.X_op != O_constant)
|
|
{
|
|
as_bad (_("Operand to .vframesp must be a constant (sp-relative offset)"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_mem_stack_v (), sep);
|
|
add_unwind_entry (output_psp_sprel (e.X_add_number), NOT_A_CHAR);
|
|
}
|
|
|
|
static void
|
|
dot_save (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e1, e2;
|
|
unsigned reg1, reg2;
|
|
int sep;
|
|
|
|
if (!in_prologue ("save"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e1, ',');
|
|
if (sep == ',')
|
|
sep = parse_operand_and_eval (&e2, ',');
|
|
else
|
|
e2.X_op = O_absent;
|
|
|
|
reg1 = e1.X_add_number;
|
|
/* Make sure it's a valid ar.xxx reg, OR its br0, aka 'rp'. */
|
|
if (e1.X_op != O_register)
|
|
{
|
|
as_bad (_("First operand to .save not a register"));
|
|
reg1 = REG_PR; /* Anything valid is good here. */
|
|
}
|
|
reg2 = e2.X_add_number - REG_GR;
|
|
if (e2.X_op != O_register || reg2 > 127)
|
|
{
|
|
as_bad (_("Second operand to .save not a valid register"));
|
|
reg2 = 0;
|
|
}
|
|
switch (reg1)
|
|
{
|
|
case REG_AR + AR_BSP:
|
|
add_unwind_entry (output_bsp_when (), sep);
|
|
add_unwind_entry (output_bsp_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_BSPSTORE:
|
|
add_unwind_entry (output_bspstore_when (), sep);
|
|
add_unwind_entry (output_bspstore_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_RNAT:
|
|
add_unwind_entry (output_rnat_when (), sep);
|
|
add_unwind_entry (output_rnat_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_UNAT:
|
|
add_unwind_entry (output_unat_when (), sep);
|
|
add_unwind_entry (output_unat_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_FPSR:
|
|
add_unwind_entry (output_fpsr_when (), sep);
|
|
add_unwind_entry (output_fpsr_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_PFS:
|
|
add_unwind_entry (output_pfs_when (), sep);
|
|
if (! (unwind.prologue_mask & 4))
|
|
add_unwind_entry (output_pfs_gr (reg2), NOT_A_CHAR);
|
|
else if (reg2 != unwind.prologue_gr
|
|
+ (unsigned) popcount (unwind.prologue_mask & -(4 << 1)))
|
|
as_warn (_("Second operand of .save contradicts .prologue"));
|
|
break;
|
|
case REG_AR + AR_LC:
|
|
add_unwind_entry (output_lc_when (), sep);
|
|
add_unwind_entry (output_lc_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
case REG_BR:
|
|
add_unwind_entry (output_rp_when (), sep);
|
|
if (! (unwind.prologue_mask & 8))
|
|
add_unwind_entry (output_rp_gr (reg2), NOT_A_CHAR);
|
|
else if (reg2 != unwind.prologue_gr)
|
|
as_warn (_("Second operand of .save contradicts .prologue"));
|
|
break;
|
|
case REG_PR:
|
|
add_unwind_entry (output_preds_when (), sep);
|
|
if (! (unwind.prologue_mask & 1))
|
|
add_unwind_entry (output_preds_gr (reg2), NOT_A_CHAR);
|
|
else if (reg2 != unwind.prologue_gr
|
|
+ (unsigned) popcount (unwind.prologue_mask & -(1 << 1)))
|
|
as_warn (_("Second operand of .save contradicts .prologue"));
|
|
break;
|
|
case REG_PRIUNAT:
|
|
add_unwind_entry (output_priunat_when_gr (), sep);
|
|
add_unwind_entry (output_priunat_gr (reg2), NOT_A_CHAR);
|
|
break;
|
|
default:
|
|
as_bad (_("First operand to .save not a valid register"));
|
|
add_unwind_entry (NULL, sep);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
dot_restore (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e1;
|
|
unsigned long ecount; /* # of _additional_ regions to pop */
|
|
int sep;
|
|
|
|
if (!in_body ("restore"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e1, ',');
|
|
if (e1.X_op != O_register || e1.X_add_number != REG_GR + 12)
|
|
as_bad (_("First operand to .restore must be stack pointer (sp)"));
|
|
|
|
if (sep == ',')
|
|
{
|
|
expressionS e2;
|
|
|
|
sep = parse_operand_and_eval (&e2, ',');
|
|
if (e2.X_op != O_constant || e2.X_add_number < 0)
|
|
{
|
|
as_bad (_("Second operand to .restore must be a constant >= 0"));
|
|
e2.X_add_number = 0;
|
|
}
|
|
ecount = e2.X_add_number;
|
|
}
|
|
else
|
|
ecount = unwind.prologue_count - 1;
|
|
|
|
if (ecount >= unwind.prologue_count)
|
|
{
|
|
as_bad (_("Epilogue count of %lu exceeds number of nested prologues (%u)"),
|
|
ecount + 1, unwind.prologue_count);
|
|
ecount = 0;
|
|
}
|
|
|
|
add_unwind_entry (output_epilogue (ecount), sep);
|
|
|
|
if (ecount < unwind.prologue_count)
|
|
unwind.prologue_count -= ecount + 1;
|
|
else
|
|
unwind.prologue_count = 0;
|
|
}
|
|
|
|
static void
|
|
dot_restorereg (int pred)
|
|
{
|
|
unsigned int qp, ab, reg;
|
|
expressionS e;
|
|
int sep;
|
|
const char * const po = pred ? "restorereg.p" : "restorereg";
|
|
|
|
if (!in_procedure (po))
|
|
return;
|
|
|
|
if (pred)
|
|
sep = parse_predicate_and_operand (&e, &qp, po);
|
|
else
|
|
{
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
qp = 0;
|
|
}
|
|
convert_expr_to_ab_reg (&e, &ab, ®, po, 1 + pred);
|
|
|
|
add_unwind_entry (output_spill_reg (ab, reg, 0, 0, qp), sep);
|
|
}
|
|
|
|
static const char *special_linkonce_name[] =
|
|
{
|
|
".gnu.linkonce.ia64unw.", ".gnu.linkonce.ia64unwi."
|
|
};
|
|
|
|
static void
|
|
start_unwind_section (const segT text_seg, int sec_index)
|
|
{
|
|
/*
|
|
Use a slightly ugly scheme to derive the unwind section names from
|
|
the text section name:
|
|
|
|
text sect. unwind table sect.
|
|
name: name: comments:
|
|
---------- ----------------- --------------------------------
|
|
.text .IA_64.unwind
|
|
.text.foo .IA_64.unwind.text.foo
|
|
.foo .IA_64.unwind.foo
|
|
.gnu.linkonce.t.foo
|
|
.gnu.linkonce.ia64unw.foo
|
|
_info .IA_64.unwind_info gas issues error message (ditto)
|
|
_infoFOO .IA_64.unwind_infoFOO gas issues error message (ditto)
|
|
|
|
This mapping is done so that:
|
|
|
|
(a) An object file with unwind info only in .text will use
|
|
unwind section names .IA_64.unwind and .IA_64.unwind_info.
|
|
This follows the letter of the ABI and also ensures backwards
|
|
compatibility with older toolchains.
|
|
|
|
(b) An object file with unwind info in multiple text sections
|
|
will use separate unwind sections for each text section.
|
|
This allows us to properly set the "sh_info" and "sh_link"
|
|
fields in SHT_IA_64_UNWIND as required by the ABI and also
|
|
lets GNU ld support programs with multiple segments
|
|
containing unwind info (as might be the case for certain
|
|
embedded applications).
|
|
|
|
(c) An error is issued if there would be a name clash.
|
|
*/
|
|
|
|
const char *text_name, *sec_text_name;
|
|
char *sec_name;
|
|
const char *prefix = special_section_name [sec_index];
|
|
const char *suffix;
|
|
|
|
sec_text_name = segment_name (text_seg);
|
|
text_name = sec_text_name;
|
|
if (strncmp (text_name, "_info", 5) == 0)
|
|
{
|
|
as_bad (_("Illegal section name `%s' (causes unwind section name clash)"),
|
|
text_name);
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
if (strcmp (text_name, ".text") == 0)
|
|
text_name = "";
|
|
|
|
/* Build the unwind section name by appending the (possibly stripped)
|
|
text section name to the unwind prefix. */
|
|
suffix = text_name;
|
|
if (strncmp (text_name, ".gnu.linkonce.t.",
|
|
sizeof (".gnu.linkonce.t.") - 1) == 0)
|
|
{
|
|
prefix = special_linkonce_name [sec_index - SPECIAL_SECTION_UNWIND];
|
|
suffix += sizeof (".gnu.linkonce.t.") - 1;
|
|
}
|
|
|
|
sec_name = concat (prefix, suffix, NULL);
|
|
|
|
/* Handle COMDAT group. */
|
|
if ((text_seg->flags & SEC_LINK_ONCE) != 0
|
|
&& (elf_section_flags (text_seg) & SHF_GROUP) != 0)
|
|
{
|
|
char *section;
|
|
const char *group_name = elf_group_name (text_seg);
|
|
|
|
if (group_name == NULL)
|
|
{
|
|
as_bad (_("Group section `%s' has no group signature"),
|
|
sec_text_name);
|
|
ignore_rest_of_line ();
|
|
free (sec_name);
|
|
return;
|
|
}
|
|
|
|
/* We have to construct a fake section directive. */
|
|
section = concat (sec_name, ",\"aG\",@progbits,", group_name, ",comdat", NULL);
|
|
set_section (section);
|
|
free (section);
|
|
}
|
|
else
|
|
{
|
|
set_section (sec_name);
|
|
bfd_set_section_flags (stdoutput, now_seg,
|
|
SEC_LOAD | SEC_ALLOC | SEC_READONLY);
|
|
}
|
|
|
|
elf_linked_to_section (now_seg) = text_seg;
|
|
free (sec_name);
|
|
}
|
|
|
|
static void
|
|
generate_unwind_image (const segT text_seg)
|
|
{
|
|
int size, pad;
|
|
unw_rec_list *list;
|
|
|
|
/* Mark the end of the unwind info, so that we can compute the size of the
|
|
last unwind region. */
|
|
add_unwind_entry (output_endp (), NOT_A_CHAR);
|
|
|
|
/* Force out pending instructions, to make sure all unwind records have
|
|
a valid slot_number field. */
|
|
ia64_flush_insns ();
|
|
|
|
/* Generate the unwind record. */
|
|
list = optimize_unw_records (unwind.list);
|
|
fixup_unw_records (list, 1);
|
|
size = calc_record_size (list);
|
|
|
|
if (size > 0 || unwind.force_unwind_entry)
|
|
{
|
|
unwind.force_unwind_entry = 0;
|
|
/* pad to pointer-size boundary. */
|
|
pad = size % md.pointer_size;
|
|
if (pad != 0)
|
|
size += md.pointer_size - pad;
|
|
/* Add 8 for the header. */
|
|
size += 8;
|
|
/* Add a pointer for the personality offset. */
|
|
if (unwind.personality_routine)
|
|
size += md.pointer_size;
|
|
}
|
|
|
|
/* If there are unwind records, switch sections, and output the info. */
|
|
if (size != 0)
|
|
{
|
|
expressionS exp;
|
|
bfd_reloc_code_real_type reloc;
|
|
|
|
start_unwind_section (text_seg, SPECIAL_SECTION_UNWIND_INFO);
|
|
|
|
/* Make sure the section has 4 byte alignment for ILP32 and
|
|
8 byte alignment for LP64. */
|
|
frag_align (md.pointer_size_shift, 0, 0);
|
|
record_alignment (now_seg, md.pointer_size_shift);
|
|
|
|
/* Set expression which points to start of unwind descriptor area. */
|
|
unwind.info = expr_build_dot ();
|
|
|
|
frag_var (rs_machine_dependent, size, size, 0, 0,
|
|
(offsetT) (long) unwind.personality_routine,
|
|
(char *) list);
|
|
|
|
/* Add the personality address to the image. */
|
|
if (unwind.personality_routine != 0)
|
|
{
|
|
exp.X_op = O_symbol;
|
|
exp.X_add_symbol = unwind.personality_routine;
|
|
exp.X_add_number = 0;
|
|
|
|
if (md.flags & EF_IA_64_BE)
|
|
{
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
reloc = BFD_RELOC_IA64_LTOFF_FPTR64MSB;
|
|
else
|
|
reloc = BFD_RELOC_IA64_LTOFF_FPTR32MSB;
|
|
}
|
|
else
|
|
{
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
reloc = BFD_RELOC_IA64_LTOFF_FPTR64LSB;
|
|
else
|
|
reloc = BFD_RELOC_IA64_LTOFF_FPTR32LSB;
|
|
}
|
|
|
|
fix_new_exp (frag_now, frag_now_fix () - md.pointer_size,
|
|
md.pointer_size, &exp, 0, reloc);
|
|
unwind.personality_routine = 0;
|
|
}
|
|
}
|
|
|
|
free_saved_prologue_counts ();
|
|
unwind.list = unwind.tail = unwind.current_entry = NULL;
|
|
}
|
|
|
|
static void
|
|
dot_handlerdata (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
if (!in_procedure ("handlerdata"))
|
|
return;
|
|
unwind.force_unwind_entry = 1;
|
|
|
|
/* Remember which segment we're in so we can switch back after .endp */
|
|
unwind.saved_text_seg = now_seg;
|
|
unwind.saved_text_subseg = now_subseg;
|
|
|
|
/* Generate unwind info into unwind-info section and then leave that
|
|
section as the currently active one so dataXX directives go into
|
|
the language specific data area of the unwind info block. */
|
|
generate_unwind_image (now_seg);
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_unwentry (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
if (!in_procedure ("unwentry"))
|
|
return;
|
|
unwind.force_unwind_entry = 1;
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_altrp (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
unsigned reg;
|
|
|
|
if (!in_prologue ("altrp"))
|
|
return;
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
reg = e.X_add_number - REG_BR;
|
|
if (e.X_op != O_register || reg > 7)
|
|
{
|
|
as_bad (_("First operand to .altrp not a valid branch register"));
|
|
reg = 0;
|
|
}
|
|
add_unwind_entry (output_rp_br (reg), 0);
|
|
}
|
|
|
|
static void
|
|
dot_savemem (int psprel)
|
|
{
|
|
expressionS e1, e2;
|
|
int sep;
|
|
int reg1, val;
|
|
const char * const po = psprel ? "savepsp" : "savesp";
|
|
|
|
if (!in_prologue (po))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e1, ',');
|
|
if (sep == ',')
|
|
sep = parse_operand_and_eval (&e2, ',');
|
|
else
|
|
e2.X_op = O_absent;
|
|
|
|
reg1 = e1.X_add_number;
|
|
val = e2.X_add_number;
|
|
|
|
/* Make sure it's a valid ar.xxx reg, OR its br0, aka 'rp'. */
|
|
if (e1.X_op != O_register)
|
|
{
|
|
as_bad (_("First operand to .%s not a register"), po);
|
|
reg1 = REG_PR; /* Anything valid is good here. */
|
|
}
|
|
if (e2.X_op != O_constant)
|
|
{
|
|
as_bad (_("Second operand to .%s not a constant"), po);
|
|
val = 0;
|
|
}
|
|
|
|
switch (reg1)
|
|
{
|
|
case REG_AR + AR_BSP:
|
|
add_unwind_entry (output_bsp_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_bsp_psprel
|
|
: output_bsp_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_BSPSTORE:
|
|
add_unwind_entry (output_bspstore_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_bspstore_psprel
|
|
: output_bspstore_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_RNAT:
|
|
add_unwind_entry (output_rnat_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_rnat_psprel
|
|
: output_rnat_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_UNAT:
|
|
add_unwind_entry (output_unat_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_unat_psprel
|
|
: output_unat_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_FPSR:
|
|
add_unwind_entry (output_fpsr_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_fpsr_psprel
|
|
: output_fpsr_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_PFS:
|
|
add_unwind_entry (output_pfs_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_pfs_psprel
|
|
: output_pfs_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_AR + AR_LC:
|
|
add_unwind_entry (output_lc_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_lc_psprel
|
|
: output_lc_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_BR:
|
|
add_unwind_entry (output_rp_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_rp_psprel
|
|
: output_rp_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_PR:
|
|
add_unwind_entry (output_preds_when (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_preds_psprel
|
|
: output_preds_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
case REG_PRIUNAT:
|
|
add_unwind_entry (output_priunat_when_mem (), sep);
|
|
add_unwind_entry ((psprel
|
|
? output_priunat_psprel
|
|
: output_priunat_sprel) (val), NOT_A_CHAR);
|
|
break;
|
|
default:
|
|
as_bad (_("First operand to .%s not a valid register"), po);
|
|
add_unwind_entry (NULL, sep);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
dot_saveg (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
unsigned grmask;
|
|
int sep;
|
|
|
|
if (!in_prologue ("save.g"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
|
|
grmask = e.X_add_number;
|
|
if (e.X_op != O_constant
|
|
|| e.X_add_number <= 0
|
|
|| e.X_add_number > 0xf)
|
|
{
|
|
as_bad (_("First operand to .save.g must be a positive 4-bit constant"));
|
|
grmask = 0;
|
|
}
|
|
|
|
if (sep == ',')
|
|
{
|
|
unsigned reg;
|
|
int n = popcount (grmask);
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
reg = e.X_add_number - REG_GR;
|
|
if (e.X_op != O_register || reg > 127)
|
|
{
|
|
as_bad (_("Second operand to .save.g must be a general register"));
|
|
reg = 0;
|
|
}
|
|
else if (reg > 128U - n)
|
|
{
|
|
as_bad (_("Second operand to .save.g must be the first of %d general registers"), n);
|
|
reg = 0;
|
|
}
|
|
add_unwind_entry (output_gr_gr (grmask, reg), 0);
|
|
}
|
|
else
|
|
add_unwind_entry (output_gr_mem (grmask), 0);
|
|
}
|
|
|
|
static void
|
|
dot_savef (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
|
|
if (!in_prologue ("save.f"))
|
|
return;
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
|
|
if (e.X_op != O_constant
|
|
|| e.X_add_number <= 0
|
|
|| e.X_add_number > 0xfffff)
|
|
{
|
|
as_bad (_("Operand to .save.f must be a positive 20-bit constant"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_fr_mem (e.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_saveb (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
unsigned brmask;
|
|
int sep;
|
|
|
|
if (!in_prologue ("save.b"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
|
|
brmask = e.X_add_number;
|
|
if (e.X_op != O_constant
|
|
|| e.X_add_number <= 0
|
|
|| e.X_add_number > 0x1f)
|
|
{
|
|
as_bad (_("First operand to .save.b must be a positive 5-bit constant"));
|
|
brmask = 0;
|
|
}
|
|
|
|
if (sep == ',')
|
|
{
|
|
unsigned reg;
|
|
int n = popcount (brmask);
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
reg = e.X_add_number - REG_GR;
|
|
if (e.X_op != O_register || reg > 127)
|
|
{
|
|
as_bad (_("Second operand to .save.b must be a general register"));
|
|
reg = 0;
|
|
}
|
|
else if (reg > 128U - n)
|
|
{
|
|
as_bad (_("Second operand to .save.b must be the first of %d general registers"), n);
|
|
reg = 0;
|
|
}
|
|
add_unwind_entry (output_br_gr (brmask, reg), 0);
|
|
}
|
|
else
|
|
add_unwind_entry (output_br_mem (brmask), 0);
|
|
}
|
|
|
|
static void
|
|
dot_savegf (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e1, e2;
|
|
|
|
if (!in_prologue ("save.gf"))
|
|
return;
|
|
|
|
if (parse_operand_and_eval (&e1, ',') == ',')
|
|
parse_operand_and_eval (&e2, 0);
|
|
else
|
|
e2.X_op = O_absent;
|
|
|
|
if (e1.X_op != O_constant
|
|
|| e1.X_add_number < 0
|
|
|| e1.X_add_number > 0xf)
|
|
{
|
|
as_bad (_("First operand to .save.gf must be a non-negative 4-bit constant"));
|
|
e1.X_op = O_absent;
|
|
e1.X_add_number = 0;
|
|
}
|
|
if (e2.X_op != O_constant
|
|
|| e2.X_add_number < 0
|
|
|| e2.X_add_number > 0xfffff)
|
|
{
|
|
as_bad (_("Second operand to .save.gf must be a non-negative 20-bit constant"));
|
|
e2.X_op = O_absent;
|
|
e2.X_add_number = 0;
|
|
}
|
|
if (e1.X_op == O_constant
|
|
&& e2.X_op == O_constant
|
|
&& e1.X_add_number == 0
|
|
&& e2.X_add_number == 0)
|
|
as_bad (_("Operands to .save.gf may not be both zero"));
|
|
|
|
add_unwind_entry (output_frgr_mem (e1.X_add_number, e2.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_spill (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
|
|
if (!in_prologue ("spill"))
|
|
return;
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
|
|
if (e.X_op != O_constant)
|
|
{
|
|
as_bad (_("Operand to .spill must be a constant"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_spill_base (e.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_spillreg (int pred)
|
|
{
|
|
int sep;
|
|
unsigned int qp, ab, xy, reg, treg;
|
|
expressionS e;
|
|
const char * const po = pred ? "spillreg.p" : "spillreg";
|
|
|
|
if (!in_procedure (po))
|
|
return;
|
|
|
|
if (pred)
|
|
sep = parse_predicate_and_operand (&e, &qp, po);
|
|
else
|
|
{
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
qp = 0;
|
|
}
|
|
convert_expr_to_ab_reg (&e, &ab, ®, po, 1 + pred);
|
|
|
|
if (sep == ',')
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
else
|
|
e.X_op = O_absent;
|
|
convert_expr_to_xy_reg (&e, &xy, &treg, po, 2 + pred);
|
|
|
|
add_unwind_entry (output_spill_reg (ab, reg, treg, xy, qp), sep);
|
|
}
|
|
|
|
static void
|
|
dot_spillmem (int psprel)
|
|
{
|
|
expressionS e;
|
|
int pred = (psprel < 0), sep;
|
|
unsigned int qp, ab, reg;
|
|
const char * po;
|
|
|
|
if (pred)
|
|
{
|
|
psprel = ~psprel;
|
|
po = psprel ? "spillpsp.p" : "spillsp.p";
|
|
}
|
|
else
|
|
po = psprel ? "spillpsp" : "spillsp";
|
|
|
|
if (!in_procedure (po))
|
|
return;
|
|
|
|
if (pred)
|
|
sep = parse_predicate_and_operand (&e, &qp, po);
|
|
else
|
|
{
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
qp = 0;
|
|
}
|
|
convert_expr_to_ab_reg (&e, &ab, ®, po, 1 + pred);
|
|
|
|
if (sep == ',')
|
|
sep = parse_operand_and_eval (&e, ',');
|
|
else
|
|
e.X_op = O_absent;
|
|
if (e.X_op != O_constant)
|
|
{
|
|
as_bad (_("Operand %d to .%s must be a constant"), 2 + pred, po);
|
|
e.X_add_number = 0;
|
|
}
|
|
|
|
if (psprel)
|
|
add_unwind_entry (output_spill_psprel (ab, reg, e.X_add_number, qp), sep);
|
|
else
|
|
add_unwind_entry (output_spill_sprel (ab, reg, e.X_add_number, qp), sep);
|
|
}
|
|
|
|
static unsigned int
|
|
get_saved_prologue_count (unsigned long lbl)
|
|
{
|
|
label_prologue_count *lpc = unwind.saved_prologue_counts;
|
|
|
|
while (lpc != NULL && lpc->label_number != lbl)
|
|
lpc = lpc->next;
|
|
|
|
if (lpc != NULL)
|
|
return lpc->prologue_count;
|
|
|
|
as_bad (_("Missing .label_state %ld"), lbl);
|
|
return 1;
|
|
}
|
|
|
|
static void
|
|
save_prologue_count (unsigned long lbl, unsigned int count)
|
|
{
|
|
label_prologue_count *lpc = unwind.saved_prologue_counts;
|
|
|
|
while (lpc != NULL && lpc->label_number != lbl)
|
|
lpc = lpc->next;
|
|
|
|
if (lpc != NULL)
|
|
lpc->prologue_count = count;
|
|
else
|
|
{
|
|
label_prologue_count *new_lpc = XNEW (label_prologue_count);
|
|
|
|
new_lpc->next = unwind.saved_prologue_counts;
|
|
new_lpc->label_number = lbl;
|
|
new_lpc->prologue_count = count;
|
|
unwind.saved_prologue_counts = new_lpc;
|
|
}
|
|
}
|
|
|
|
static void
|
|
free_saved_prologue_counts (void)
|
|
{
|
|
label_prologue_count *lpc = unwind.saved_prologue_counts;
|
|
label_prologue_count *next;
|
|
|
|
while (lpc != NULL)
|
|
{
|
|
next = lpc->next;
|
|
free (lpc);
|
|
lpc = next;
|
|
}
|
|
|
|
unwind.saved_prologue_counts = NULL;
|
|
}
|
|
|
|
static void
|
|
dot_label_state (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
|
|
if (!in_body ("label_state"))
|
|
return;
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
if (e.X_op == O_constant)
|
|
save_prologue_count (e.X_add_number, unwind.prologue_count);
|
|
else
|
|
{
|
|
as_bad (_("Operand to .label_state must be a constant"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_label_state (e.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_copy_state (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
|
|
if (!in_body ("copy_state"))
|
|
return;
|
|
|
|
parse_operand_and_eval (&e, 0);
|
|
if (e.X_op == O_constant)
|
|
unwind.prologue_count = get_saved_prologue_count (e.X_add_number);
|
|
else
|
|
{
|
|
as_bad (_("Operand to .copy_state must be a constant"));
|
|
e.X_add_number = 0;
|
|
}
|
|
add_unwind_entry (output_copy_state (e.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_unwabi (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e1, e2;
|
|
unsigned char sep;
|
|
|
|
if (!in_prologue ("unwabi"))
|
|
return;
|
|
|
|
sep = parse_operand_and_eval (&e1, ',');
|
|
if (sep == ',')
|
|
parse_operand_and_eval (&e2, 0);
|
|
else
|
|
e2.X_op = O_absent;
|
|
|
|
if (e1.X_op != O_constant)
|
|
{
|
|
as_bad (_("First operand to .unwabi must be a constant"));
|
|
e1.X_add_number = 0;
|
|
}
|
|
|
|
if (e2.X_op != O_constant)
|
|
{
|
|
as_bad (_("Second operand to .unwabi must be a constant"));
|
|
e2.X_add_number = 0;
|
|
}
|
|
|
|
add_unwind_entry (output_unwabi (e1.X_add_number, e2.X_add_number), 0);
|
|
}
|
|
|
|
static void
|
|
dot_personality (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
char *name, *p, c;
|
|
|
|
if (!in_procedure ("personality"))
|
|
return;
|
|
SKIP_WHITESPACE ();
|
|
c = get_symbol_name (&name);
|
|
p = input_line_pointer;
|
|
unwind.personality_routine = symbol_find_or_make (name);
|
|
unwind.force_unwind_entry = 1;
|
|
*p = c;
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_proc (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
char *name, *p, c;
|
|
symbolS *sym;
|
|
proc_pending *pending, *last_pending;
|
|
|
|
if (unwind.proc_pending.sym)
|
|
{
|
|
(md.unwind_check == unwind_check_warning
|
|
? as_warn
|
|
: as_bad) (_("Missing .endp after previous .proc"));
|
|
while (unwind.proc_pending.next)
|
|
{
|
|
pending = unwind.proc_pending.next;
|
|
unwind.proc_pending.next = pending->next;
|
|
free (pending);
|
|
}
|
|
}
|
|
last_pending = NULL;
|
|
|
|
/* Parse names of main and alternate entry points and mark them as
|
|
function symbols: */
|
|
while (1)
|
|
{
|
|
SKIP_WHITESPACE ();
|
|
c = get_symbol_name (&name);
|
|
p = input_line_pointer;
|
|
if (!*name)
|
|
as_bad (_("Empty argument of .proc"));
|
|
else
|
|
{
|
|
sym = symbol_find_or_make (name);
|
|
if (S_IS_DEFINED (sym))
|
|
as_bad (_("`%s' was already defined"), name);
|
|
else if (!last_pending)
|
|
{
|
|
unwind.proc_pending.sym = sym;
|
|
last_pending = &unwind.proc_pending;
|
|
}
|
|
else
|
|
{
|
|
pending = XNEW (proc_pending);
|
|
pending->sym = sym;
|
|
last_pending = last_pending->next = pending;
|
|
}
|
|
symbol_get_bfdsym (sym)->flags |= BSF_FUNCTION;
|
|
}
|
|
*p = c;
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
if (*input_line_pointer != ',')
|
|
break;
|
|
++input_line_pointer;
|
|
}
|
|
if (!last_pending)
|
|
{
|
|
unwind.proc_pending.sym = expr_build_dot ();
|
|
last_pending = &unwind.proc_pending;
|
|
}
|
|
last_pending->next = NULL;
|
|
demand_empty_rest_of_line ();
|
|
do_align (4, NULL, 0, 0);
|
|
|
|
unwind.prologue = 0;
|
|
unwind.prologue_count = 0;
|
|
unwind.body = 0;
|
|
unwind.insn = 0;
|
|
unwind.list = unwind.tail = unwind.current_entry = NULL;
|
|
unwind.personality_routine = 0;
|
|
}
|
|
|
|
static void
|
|
dot_body (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
if (!in_procedure ("body"))
|
|
return;
|
|
if (!unwind.prologue && !unwind.body && unwind.insn)
|
|
as_warn (_("Initial .body should precede any instructions"));
|
|
check_pending_save ();
|
|
|
|
unwind.prologue = 0;
|
|
unwind.prologue_mask = 0;
|
|
unwind.body = 1;
|
|
|
|
add_unwind_entry (output_body (), 0);
|
|
}
|
|
|
|
static void
|
|
dot_prologue (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
unsigned mask = 0, grsave = 0;
|
|
|
|
if (!in_procedure ("prologue"))
|
|
return;
|
|
if (unwind.prologue)
|
|
{
|
|
as_bad (_(".prologue within prologue"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
if (!unwind.body && unwind.insn)
|
|
as_warn (_("Initial .prologue should precede any instructions"));
|
|
|
|
if (!is_it_end_of_statement ())
|
|
{
|
|
expressionS e;
|
|
int n, sep = parse_operand_and_eval (&e, ',');
|
|
|
|
if (e.X_op != O_constant
|
|
|| e.X_add_number < 0
|
|
|| e.X_add_number > 0xf)
|
|
as_bad (_("First operand to .prologue must be a positive 4-bit constant"));
|
|
else if (e.X_add_number == 0)
|
|
as_warn (_("Pointless use of zero first operand to .prologue"));
|
|
else
|
|
mask = e.X_add_number;
|
|
|
|
n = popcount (mask);
|
|
|
|
if (sep == ',')
|
|
parse_operand_and_eval (&e, 0);
|
|
else
|
|
e.X_op = O_absent;
|
|
|
|
if (e.X_op == O_constant
|
|
&& e.X_add_number >= 0
|
|
&& e.X_add_number < 128)
|
|
{
|
|
if (md.unwind_check == unwind_check_error)
|
|
as_warn (_("Using a constant as second operand to .prologue is deprecated"));
|
|
grsave = e.X_add_number;
|
|
}
|
|
else if (e.X_op != O_register
|
|
|| (grsave = e.X_add_number - REG_GR) > 127)
|
|
{
|
|
as_bad (_("Second operand to .prologue must be a general register"));
|
|
grsave = 0;
|
|
}
|
|
else if (grsave > 128U - n)
|
|
{
|
|
as_bad (_("Second operand to .prologue must be the first of %d general registers"), n);
|
|
grsave = 0;
|
|
}
|
|
}
|
|
|
|
if (mask)
|
|
add_unwind_entry (output_prologue_gr (mask, grsave), 0);
|
|
else
|
|
add_unwind_entry (output_prologue (), 0);
|
|
|
|
unwind.prologue = 1;
|
|
unwind.prologue_mask = mask;
|
|
unwind.prologue_gr = grsave;
|
|
unwind.body = 0;
|
|
++unwind.prologue_count;
|
|
}
|
|
|
|
static void
|
|
dot_endp (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS e;
|
|
int bytes_per_address;
|
|
long where;
|
|
segT saved_seg;
|
|
subsegT saved_subseg;
|
|
proc_pending *pending;
|
|
int unwind_check = md.unwind_check;
|
|
|
|
md.unwind_check = unwind_check_error;
|
|
if (!in_procedure ("endp"))
|
|
return;
|
|
md.unwind_check = unwind_check;
|
|
|
|
if (unwind.saved_text_seg)
|
|
{
|
|
saved_seg = unwind.saved_text_seg;
|
|
saved_subseg = unwind.saved_text_subseg;
|
|
unwind.saved_text_seg = NULL;
|
|
}
|
|
else
|
|
{
|
|
saved_seg = now_seg;
|
|
saved_subseg = now_subseg;
|
|
}
|
|
|
|
insn_group_break (1, 0, 0);
|
|
|
|
/* If there wasn't a .handlerdata, we haven't generated an image yet. */
|
|
if (!unwind.info)
|
|
generate_unwind_image (saved_seg);
|
|
|
|
if (unwind.info || unwind.force_unwind_entry)
|
|
{
|
|
symbolS *proc_end;
|
|
|
|
subseg_set (md.last_text_seg, 0);
|
|
proc_end = expr_build_dot ();
|
|
|
|
start_unwind_section (saved_seg, SPECIAL_SECTION_UNWIND);
|
|
|
|
/* Make sure that section has 4 byte alignment for ILP32 and
|
|
8 byte alignment for LP64. */
|
|
record_alignment (now_seg, md.pointer_size_shift);
|
|
|
|
/* Need space for 3 pointers for procedure start, procedure end,
|
|
and unwind info. */
|
|
memset (frag_more (3 * md.pointer_size), 0, 3 * md.pointer_size);
|
|
where = frag_now_fix () - (3 * md.pointer_size);
|
|
bytes_per_address = bfd_arch_bits_per_address (stdoutput) / 8;
|
|
|
|
/* Issue the values of a) Proc Begin, b) Proc End, c) Unwind Record. */
|
|
e.X_op = O_pseudo_fixup;
|
|
e.X_op_symbol = pseudo_func[FUNC_SEG_RELATIVE].u.sym;
|
|
e.X_add_number = 0;
|
|
if (!S_IS_LOCAL (unwind.proc_pending.sym)
|
|
&& S_IS_DEFINED (unwind.proc_pending.sym))
|
|
e.X_add_symbol = symbol_temp_new (S_GET_SEGMENT (unwind.proc_pending.sym),
|
|
S_GET_VALUE (unwind.proc_pending.sym),
|
|
symbol_get_frag (unwind.proc_pending.sym));
|
|
else
|
|
e.X_add_symbol = unwind.proc_pending.sym;
|
|
ia64_cons_fix_new (frag_now, where, bytes_per_address, &e,
|
|
BFD_RELOC_NONE);
|
|
|
|
e.X_op = O_pseudo_fixup;
|
|
e.X_op_symbol = pseudo_func[FUNC_SEG_RELATIVE].u.sym;
|
|
e.X_add_number = 0;
|
|
e.X_add_symbol = proc_end;
|
|
ia64_cons_fix_new (frag_now, where + bytes_per_address,
|
|
bytes_per_address, &e, BFD_RELOC_NONE);
|
|
|
|
if (unwind.info)
|
|
{
|
|
e.X_op = O_pseudo_fixup;
|
|
e.X_op_symbol = pseudo_func[FUNC_SEG_RELATIVE].u.sym;
|
|
e.X_add_number = 0;
|
|
e.X_add_symbol = unwind.info;
|
|
ia64_cons_fix_new (frag_now, where + (bytes_per_address * 2),
|
|
bytes_per_address, &e, BFD_RELOC_NONE);
|
|
}
|
|
}
|
|
subseg_set (saved_seg, saved_subseg);
|
|
|
|
/* Set symbol sizes. */
|
|
pending = &unwind.proc_pending;
|
|
if (S_GET_NAME (pending->sym))
|
|
{
|
|
do
|
|
{
|
|
symbolS *sym = pending->sym;
|
|
|
|
if (!S_IS_DEFINED (sym))
|
|
as_bad (_("`%s' was not defined within procedure"), S_GET_NAME (sym));
|
|
else if (S_GET_SIZE (sym) == 0
|
|
&& symbol_get_obj (sym)->size == NULL)
|
|
{
|
|
fragS *frag = symbol_get_frag (sym);
|
|
|
|
if (frag)
|
|
{
|
|
if (frag == frag_now && SEG_NORMAL (now_seg))
|
|
S_SET_SIZE (sym, frag_now_fix () - S_GET_VALUE (sym));
|
|
else
|
|
{
|
|
symbol_get_obj (sym)->size = XNEW (expressionS);
|
|
symbol_get_obj (sym)->size->X_op = O_subtract;
|
|
symbol_get_obj (sym)->size->X_add_symbol
|
|
= symbol_new (FAKE_LABEL_NAME, now_seg,
|
|
frag_now_fix (), frag_now);
|
|
symbol_get_obj (sym)->size->X_op_symbol = sym;
|
|
symbol_get_obj (sym)->size->X_add_number = 0;
|
|
}
|
|
}
|
|
}
|
|
} while ((pending = pending->next) != NULL);
|
|
}
|
|
|
|
/* Parse names of main and alternate entry points. */
|
|
while (1)
|
|
{
|
|
char *name, *p, c;
|
|
|
|
SKIP_WHITESPACE ();
|
|
c = get_symbol_name (&name);
|
|
p = input_line_pointer;
|
|
if (!*name)
|
|
(md.unwind_check == unwind_check_warning
|
|
? as_warn
|
|
: as_bad) (_("Empty argument of .endp"));
|
|
else
|
|
{
|
|
symbolS *sym = symbol_find (name);
|
|
|
|
for (pending = &unwind.proc_pending; pending; pending = pending->next)
|
|
{
|
|
if (sym == pending->sym)
|
|
{
|
|
pending->sym = NULL;
|
|
break;
|
|
}
|
|
}
|
|
if (!sym || !pending)
|
|
as_warn (_("`%s' was not specified with previous .proc"), name);
|
|
}
|
|
*p = c;
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
if (*input_line_pointer != ',')
|
|
break;
|
|
++input_line_pointer;
|
|
}
|
|
demand_empty_rest_of_line ();
|
|
|
|
/* Deliberately only checking for the main entry point here; the
|
|
language spec even says all arguments to .endp are ignored. */
|
|
if (unwind.proc_pending.sym
|
|
&& S_GET_NAME (unwind.proc_pending.sym)
|
|
&& strcmp (S_GET_NAME (unwind.proc_pending.sym), FAKE_LABEL_NAME))
|
|
as_warn (_("`%s' should be an operand to this .endp"),
|
|
S_GET_NAME (unwind.proc_pending.sym));
|
|
while (unwind.proc_pending.next)
|
|
{
|
|
pending = unwind.proc_pending.next;
|
|
unwind.proc_pending.next = pending->next;
|
|
free (pending);
|
|
}
|
|
unwind.proc_pending.sym = unwind.info = NULL;
|
|
}
|
|
|
|
static void
|
|
dot_template (int template_val)
|
|
{
|
|
CURR_SLOT.user_template = template_val;
|
|
}
|
|
|
|
static void
|
|
dot_regstk (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
int ins, locs, outs, rots;
|
|
|
|
if (is_it_end_of_statement ())
|
|
ins = locs = outs = rots = 0;
|
|
else
|
|
{
|
|
ins = get_absolute_expression ();
|
|
if (*input_line_pointer++ != ',')
|
|
goto err;
|
|
locs = get_absolute_expression ();
|
|
if (*input_line_pointer++ != ',')
|
|
goto err;
|
|
outs = get_absolute_expression ();
|
|
if (*input_line_pointer++ != ',')
|
|
goto err;
|
|
rots = get_absolute_expression ();
|
|
}
|
|
set_regstack (ins, locs, outs, rots);
|
|
return;
|
|
|
|
err:
|
|
as_bad (_("Comma expected"));
|
|
ignore_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_rot (int type)
|
|
{
|
|
offsetT num_regs;
|
|
valueT num_alloced = 0;
|
|
struct dynreg **drpp, *dr;
|
|
int ch, base_reg = 0;
|
|
char *name, *start;
|
|
size_t len;
|
|
|
|
switch (type)
|
|
{
|
|
case DYNREG_GR: base_reg = REG_GR + 32; break;
|
|
case DYNREG_FR: base_reg = REG_FR + 32; break;
|
|
case DYNREG_PR: base_reg = REG_P + 16; break;
|
|
default: break;
|
|
}
|
|
|
|
/* First, remove existing names from hash table. */
|
|
for (dr = md.dynreg[type]; dr && dr->num_regs; dr = dr->next)
|
|
{
|
|
hash_delete (md.dynreg_hash, dr->name, FALSE);
|
|
/* FIXME: Free dr->name. */
|
|
dr->num_regs = 0;
|
|
}
|
|
|
|
drpp = &md.dynreg[type];
|
|
while (1)
|
|
{
|
|
ch = get_symbol_name (&start);
|
|
len = strlen (ia64_canonicalize_symbol_name (start));
|
|
*input_line_pointer = ch;
|
|
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
if (*input_line_pointer != '[')
|
|
{
|
|
as_bad (_("Expected '['"));
|
|
goto err;
|
|
}
|
|
++input_line_pointer; /* skip '[' */
|
|
|
|
num_regs = get_absolute_expression ();
|
|
|
|
if (*input_line_pointer++ != ']')
|
|
{
|
|
as_bad (_("Expected ']'"));
|
|
goto err;
|
|
}
|
|
if (num_regs <= 0)
|
|
{
|
|
as_bad (_("Number of elements must be positive"));
|
|
goto err;
|
|
}
|
|
SKIP_WHITESPACE ();
|
|
|
|
num_alloced += num_regs;
|
|
switch (type)
|
|
{
|
|
case DYNREG_GR:
|
|
if (num_alloced > md.rot.num_regs)
|
|
{
|
|
as_bad (_("Used more than the declared %d rotating registers"),
|
|
md.rot.num_regs);
|
|
goto err;
|
|
}
|
|
break;
|
|
case DYNREG_FR:
|
|
if (num_alloced > 96)
|
|
{
|
|
as_bad (_("Used more than the available 96 rotating registers"));
|
|
goto err;
|
|
}
|
|
break;
|
|
case DYNREG_PR:
|
|
if (num_alloced > 48)
|
|
{
|
|
as_bad (_("Used more than the available 48 rotating registers"));
|
|
goto err;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (!*drpp)
|
|
{
|
|
*drpp = XOBNEW (¬es, struct dynreg);
|
|
memset (*drpp, 0, sizeof (*dr));
|
|
}
|
|
|
|
name = XOBNEWVEC (¬es, char, len + 1);
|
|
memcpy (name, start, len);
|
|
name[len] = '\0';
|
|
|
|
dr = *drpp;
|
|
dr->name = name;
|
|
dr->num_regs = num_regs;
|
|
dr->base = base_reg;
|
|
drpp = &dr->next;
|
|
base_reg += num_regs;
|
|
|
|
if (hash_insert (md.dynreg_hash, name, dr))
|
|
{
|
|
as_bad (_("Attempt to redefine register set `%s'"), name);
|
|
obstack_free (¬es, name);
|
|
goto err;
|
|
}
|
|
|
|
if (*input_line_pointer != ',')
|
|
break;
|
|
++input_line_pointer; /* skip comma */
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
demand_empty_rest_of_line ();
|
|
return;
|
|
|
|
err:
|
|
ignore_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_byteorder (int byteorder)
|
|
{
|
|
segment_info_type *seginfo = seg_info (now_seg);
|
|
|
|
if (byteorder == -1)
|
|
{
|
|
if (seginfo->tc_segment_info_data.endian == 0)
|
|
seginfo->tc_segment_info_data.endian = default_big_endian ? 1 : 2;
|
|
byteorder = seginfo->tc_segment_info_data.endian == 1;
|
|
}
|
|
else
|
|
seginfo->tc_segment_info_data.endian = byteorder ? 1 : 2;
|
|
|
|
if (target_big_endian != byteorder)
|
|
{
|
|
target_big_endian = byteorder;
|
|
if (target_big_endian)
|
|
{
|
|
ia64_number_to_chars = number_to_chars_bigendian;
|
|
ia64_float_to_chars = ia64_float_to_chars_bigendian;
|
|
}
|
|
else
|
|
{
|
|
ia64_number_to_chars = number_to_chars_littleendian;
|
|
ia64_float_to_chars = ia64_float_to_chars_littleendian;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
dot_psr (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
char *option;
|
|
int ch;
|
|
|
|
while (1)
|
|
{
|
|
ch = get_symbol_name (&option);
|
|
if (strcmp (option, "lsb") == 0)
|
|
md.flags &= ~EF_IA_64_BE;
|
|
else if (strcmp (option, "msb") == 0)
|
|
md.flags |= EF_IA_64_BE;
|
|
else if (strcmp (option, "abi32") == 0)
|
|
md.flags &= ~EF_IA_64_ABI64;
|
|
else if (strcmp (option, "abi64") == 0)
|
|
md.flags |= EF_IA_64_ABI64;
|
|
else
|
|
as_bad (_("Unknown psr option `%s'"), option);
|
|
*input_line_pointer = ch;
|
|
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
if (*input_line_pointer != ',')
|
|
break;
|
|
|
|
++input_line_pointer;
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
dot_ln (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
new_logical_line (0, get_absolute_expression ());
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
static void
|
|
cross_section (int ref, void (*builder) (int), int ua)
|
|
{
|
|
char *start, *end;
|
|
int saved_auto_align;
|
|
unsigned int section_count;
|
|
char *name;
|
|
char c;
|
|
|
|
SKIP_WHITESPACE ();
|
|
start = input_line_pointer;
|
|
c = get_symbol_name (&name);
|
|
if (input_line_pointer == start)
|
|
{
|
|
as_bad (_("Missing section name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
* input_line_pointer = c;
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
end = input_line_pointer;
|
|
if (*input_line_pointer != ',')
|
|
{
|
|
as_bad (_("Comma expected after section name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
*end = '\0';
|
|
end = input_line_pointer + 1; /* skip comma */
|
|
input_line_pointer = start;
|
|
md.keep_pending_output = 1;
|
|
section_count = bfd_count_sections (stdoutput);
|
|
obj_elf_section (0);
|
|
if (section_count != bfd_count_sections (stdoutput))
|
|
as_warn (_("Creating sections with .xdataN/.xrealN/.xstringZ is deprecated."));
|
|
input_line_pointer = end;
|
|
saved_auto_align = md.auto_align;
|
|
if (ua)
|
|
md.auto_align = 0;
|
|
(*builder) (ref);
|
|
if (ua)
|
|
md.auto_align = saved_auto_align;
|
|
obj_elf_previous (0);
|
|
md.keep_pending_output = 0;
|
|
}
|
|
|
|
static void
|
|
dot_xdata (int size)
|
|
{
|
|
cross_section (size, cons, 0);
|
|
}
|
|
|
|
/* Why doesn't float_cons() call md_cons_align() the way cons() does? */
|
|
|
|
static void
|
|
stmt_float_cons (int kind)
|
|
{
|
|
size_t alignment;
|
|
|
|
switch (kind)
|
|
{
|
|
case 'd':
|
|
alignment = 3;
|
|
break;
|
|
|
|
case 'x':
|
|
case 'X':
|
|
alignment = 4;
|
|
break;
|
|
|
|
case 'f':
|
|
default:
|
|
alignment = 2;
|
|
break;
|
|
}
|
|
do_align (alignment, NULL, 0, 0);
|
|
float_cons (kind);
|
|
}
|
|
|
|
static void
|
|
stmt_cons_ua (int size)
|
|
{
|
|
int saved_auto_align = md.auto_align;
|
|
|
|
md.auto_align = 0;
|
|
cons (size);
|
|
md.auto_align = saved_auto_align;
|
|
}
|
|
|
|
static void
|
|
dot_xfloat_cons (int kind)
|
|
{
|
|
cross_section (kind, stmt_float_cons, 0);
|
|
}
|
|
|
|
static void
|
|
dot_xstringer (int zero)
|
|
{
|
|
cross_section (zero, stringer, 0);
|
|
}
|
|
|
|
static void
|
|
dot_xdata_ua (int size)
|
|
{
|
|
cross_section (size, cons, 1);
|
|
}
|
|
|
|
static void
|
|
dot_xfloat_cons_ua (int kind)
|
|
{
|
|
cross_section (kind, float_cons, 1);
|
|
}
|
|
|
|
/* .reg.val <regname>,value */
|
|
|
|
static void
|
|
dot_reg_val (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
expressionS reg;
|
|
|
|
expression_and_evaluate (®);
|
|
if (reg.X_op != O_register)
|
|
{
|
|
as_bad (_("Register name expected"));
|
|
ignore_rest_of_line ();
|
|
}
|
|
else if (*input_line_pointer++ != ',')
|
|
{
|
|
as_bad (_("Comma expected"));
|
|
ignore_rest_of_line ();
|
|
}
|
|
else
|
|
{
|
|
valueT value = get_absolute_expression ();
|
|
int regno = reg.X_add_number;
|
|
if (regno <= REG_GR || regno > REG_GR + 127)
|
|
as_warn (_("Register value annotation ignored"));
|
|
else
|
|
{
|
|
gr_values[regno - REG_GR].known = 1;
|
|
gr_values[regno - REG_GR].value = value;
|
|
gr_values[regno - REG_GR].path = md.path;
|
|
}
|
|
}
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/*
|
|
.serialize.data
|
|
.serialize.instruction
|
|
*/
|
|
static void
|
|
dot_serialize (int type)
|
|
{
|
|
insn_group_break (0, 0, 0);
|
|
if (type)
|
|
instruction_serialization ();
|
|
else
|
|
data_serialization ();
|
|
insn_group_break (0, 0, 0);
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* select dv checking mode
|
|
.auto
|
|
.explicit
|
|
.default
|
|
|
|
A stop is inserted when changing modes
|
|
*/
|
|
|
|
static void
|
|
dot_dv_mode (int type)
|
|
{
|
|
if (md.manual_bundling)
|
|
as_warn (_("Directive invalid within a bundle"));
|
|
|
|
if (type == 'E' || type == 'A')
|
|
md.mode_explicitly_set = 0;
|
|
else
|
|
md.mode_explicitly_set = 1;
|
|
|
|
md.detect_dv = 1;
|
|
switch (type)
|
|
{
|
|
case 'A':
|
|
case 'a':
|
|
if (md.explicit_mode)
|
|
insn_group_break (1, 0, 0);
|
|
md.explicit_mode = 0;
|
|
break;
|
|
case 'E':
|
|
case 'e':
|
|
if (!md.explicit_mode)
|
|
insn_group_break (1, 0, 0);
|
|
md.explicit_mode = 1;
|
|
break;
|
|
default:
|
|
case 'd':
|
|
if (md.explicit_mode != md.default_explicit_mode)
|
|
insn_group_break (1, 0, 0);
|
|
md.explicit_mode = md.default_explicit_mode;
|
|
md.mode_explicitly_set = 0;
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
print_prmask (valueT mask)
|
|
{
|
|
int regno;
|
|
const char *comma = "";
|
|
for (regno = 0; regno < 64; regno++)
|
|
{
|
|
if (mask & ((valueT) 1 << regno))
|
|
{
|
|
fprintf (stderr, "%s p%d", comma, regno);
|
|
comma = ",";
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
.pred.rel.clear [p1 [,p2 [,...]]] (also .pred.rel "clear" or @clear)
|
|
.pred.rel.imply p1, p2 (also .pred.rel "imply" or @imply)
|
|
.pred.rel.mutex p1, p2 [,...] (also .pred.rel "mutex" or @mutex)
|
|
.pred.safe_across_calls p1 [, p2 [,...]]
|
|
*/
|
|
|
|
static void
|
|
dot_pred_rel (int type)
|
|
{
|
|
valueT mask = 0;
|
|
int count = 0;
|
|
int p1 = -1, p2 = -1;
|
|
|
|
if (type == 0)
|
|
{
|
|
if (*input_line_pointer == '"')
|
|
{
|
|
int len;
|
|
char *form = demand_copy_C_string (&len);
|
|
|
|
if (strcmp (form, "mutex") == 0)
|
|
type = 'm';
|
|
else if (strcmp (form, "clear") == 0)
|
|
type = 'c';
|
|
else if (strcmp (form, "imply") == 0)
|
|
type = 'i';
|
|
obstack_free (¬es, form);
|
|
}
|
|
else if (*input_line_pointer == '@')
|
|
{
|
|
char *form;
|
|
char c;
|
|
|
|
++input_line_pointer;
|
|
c = get_symbol_name (&form);
|
|
|
|
if (strcmp (form, "mutex") == 0)
|
|
type = 'm';
|
|
else if (strcmp (form, "clear") == 0)
|
|
type = 'c';
|
|
else if (strcmp (form, "imply") == 0)
|
|
type = 'i';
|
|
(void) restore_line_pointer (c);
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("Missing predicate relation type"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
if (type == 0)
|
|
{
|
|
as_bad (_("Unrecognized predicate relation type"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
if (*input_line_pointer == ',')
|
|
++input_line_pointer;
|
|
SKIP_WHITESPACE ();
|
|
}
|
|
|
|
while (1)
|
|
{
|
|
valueT bits = 1;
|
|
int sep, regno;
|
|
expressionS pr, *pr1, *pr2;
|
|
|
|
sep = parse_operand_and_eval (&pr, ',');
|
|
if (pr.X_op == O_register
|
|
&& pr.X_add_number >= REG_P
|
|
&& pr.X_add_number <= REG_P + 63)
|
|
{
|
|
regno = pr.X_add_number - REG_P;
|
|
bits <<= regno;
|
|
count++;
|
|
if (p1 == -1)
|
|
p1 = regno;
|
|
else if (p2 == -1)
|
|
p2 = regno;
|
|
}
|
|
else if (type != 'i'
|
|
&& pr.X_op == O_subtract
|
|
&& (pr1 = symbol_get_value_expression (pr.X_add_symbol))
|
|
&& pr1->X_op == O_register
|
|
&& pr1->X_add_number >= REG_P
|
|
&& pr1->X_add_number <= REG_P + 63
|
|
&& (pr2 = symbol_get_value_expression (pr.X_op_symbol))
|
|
&& pr2->X_op == O_register
|
|
&& pr2->X_add_number >= REG_P
|
|
&& pr2->X_add_number <= REG_P + 63)
|
|
{
|
|
/* It's a range. */
|
|
int stop;
|
|
|
|
regno = pr1->X_add_number - REG_P;
|
|
stop = pr2->X_add_number - REG_P;
|
|
if (regno >= stop)
|
|
{
|
|
as_bad (_("Bad register range"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
bits = ((bits << stop) << 1) - (bits << regno);
|
|
count += stop - regno + 1;
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("Predicate register expected"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
if (mask & bits)
|
|
as_warn (_("Duplicate predicate register ignored"));
|
|
mask |= bits;
|
|
if (sep != ',')
|
|
break;
|
|
}
|
|
|
|
switch (type)
|
|
{
|
|
case 'c':
|
|
if (count == 0)
|
|
mask = ~(valueT) 0;
|
|
clear_qp_mutex (mask);
|
|
clear_qp_implies (mask, (valueT) 0);
|
|
break;
|
|
case 'i':
|
|
if (count != 2 || p1 == -1 || p2 == -1)
|
|
as_bad (_("Predicate source and target required"));
|
|
else if (p1 == 0 || p2 == 0)
|
|
as_bad (_("Use of p0 is not valid in this context"));
|
|
else
|
|
add_qp_imply (p1, p2);
|
|
break;
|
|
case 'm':
|
|
if (count < 2)
|
|
{
|
|
as_bad (_("At least two PR arguments expected"));
|
|
break;
|
|
}
|
|
else if (mask & 1)
|
|
{
|
|
as_bad (_("Use of p0 is not valid in this context"));
|
|
break;
|
|
}
|
|
add_qp_mutex (mask);
|
|
break;
|
|
case 's':
|
|
/* note that we don't override any existing relations */
|
|
if (count == 0)
|
|
{
|
|
as_bad (_("At least one PR argument expected"));
|
|
break;
|
|
}
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, "Safe across calls: ");
|
|
print_prmask (mask);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
qp_safe_across_calls = mask;
|
|
break;
|
|
}
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* .entry label [, label [, ...]]
|
|
Hint to DV code that the given labels are to be considered entry points.
|
|
Otherwise, only global labels are considered entry points. */
|
|
|
|
static void
|
|
dot_entry (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
const char *err;
|
|
char *name;
|
|
int c;
|
|
symbolS *symbolP;
|
|
|
|
do
|
|
{
|
|
c = get_symbol_name (&name);
|
|
symbolP = symbol_find_or_make (name);
|
|
|
|
err = hash_insert (md.entry_hash, S_GET_NAME (symbolP), (void *) symbolP);
|
|
if (err)
|
|
as_fatal (_("Inserting \"%s\" into entry hint table failed: %s"),
|
|
name, err);
|
|
|
|
*input_line_pointer = c;
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
c = *input_line_pointer;
|
|
if (c == ',')
|
|
{
|
|
input_line_pointer++;
|
|
SKIP_WHITESPACE ();
|
|
if (*input_line_pointer == '\n')
|
|
c = '\n';
|
|
}
|
|
}
|
|
while (c == ',');
|
|
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* .mem.offset offset, base
|
|
"base" is used to distinguish between offsets from a different base. */
|
|
|
|
static void
|
|
dot_mem_offset (int dummy ATTRIBUTE_UNUSED)
|
|
{
|
|
md.mem_offset.hint = 1;
|
|
md.mem_offset.offset = get_absolute_expression ();
|
|
if (*input_line_pointer != ',')
|
|
{
|
|
as_bad (_("Comma expected"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
++input_line_pointer;
|
|
md.mem_offset.base = get_absolute_expression ();
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* ia64-specific pseudo-ops: */
|
|
const pseudo_typeS md_pseudo_table[] =
|
|
{
|
|
{ "radix", dot_radix, 0 },
|
|
{ "lcomm", s_lcomm_bytes, 1 },
|
|
{ "loc", dot_loc, 0 },
|
|
{ "bss", dot_special_section, SPECIAL_SECTION_BSS },
|
|
{ "sbss", dot_special_section, SPECIAL_SECTION_SBSS },
|
|
{ "sdata", dot_special_section, SPECIAL_SECTION_SDATA },
|
|
{ "rodata", dot_special_section, SPECIAL_SECTION_RODATA },
|
|
{ "comment", dot_special_section, SPECIAL_SECTION_COMMENT },
|
|
{ "ia_64.unwind", dot_special_section, SPECIAL_SECTION_UNWIND },
|
|
{ "ia_64.unwind_info", dot_special_section, SPECIAL_SECTION_UNWIND_INFO },
|
|
{ "init_array", dot_special_section, SPECIAL_SECTION_INIT_ARRAY },
|
|
{ "fini_array", dot_special_section, SPECIAL_SECTION_FINI_ARRAY },
|
|
{ "proc", dot_proc, 0 },
|
|
{ "body", dot_body, 0 },
|
|
{ "prologue", dot_prologue, 0 },
|
|
{ "endp", dot_endp, 0 },
|
|
|
|
{ "fframe", dot_fframe, 0 },
|
|
{ "vframe", dot_vframe, 0 },
|
|
{ "vframesp", dot_vframesp, 0 },
|
|
{ "vframepsp", dot_vframesp, 1 },
|
|
{ "save", dot_save, 0 },
|
|
{ "restore", dot_restore, 0 },
|
|
{ "restorereg", dot_restorereg, 0 },
|
|
{ "restorereg.p", dot_restorereg, 1 },
|
|
{ "handlerdata", dot_handlerdata, 0 },
|
|
{ "unwentry", dot_unwentry, 0 },
|
|
{ "altrp", dot_altrp, 0 },
|
|
{ "savesp", dot_savemem, 0 },
|
|
{ "savepsp", dot_savemem, 1 },
|
|
{ "save.g", dot_saveg, 0 },
|
|
{ "save.f", dot_savef, 0 },
|
|
{ "save.b", dot_saveb, 0 },
|
|
{ "save.gf", dot_savegf, 0 },
|
|
{ "spill", dot_spill, 0 },
|
|
{ "spillreg", dot_spillreg, 0 },
|
|
{ "spillsp", dot_spillmem, 0 },
|
|
{ "spillpsp", dot_spillmem, 1 },
|
|
{ "spillreg.p", dot_spillreg, 1 },
|
|
{ "spillsp.p", dot_spillmem, ~0 },
|
|
{ "spillpsp.p", dot_spillmem, ~1 },
|
|
{ "label_state", dot_label_state, 0 },
|
|
{ "copy_state", dot_copy_state, 0 },
|
|
{ "unwabi", dot_unwabi, 0 },
|
|
{ "personality", dot_personality, 0 },
|
|
{ "mii", dot_template, 0x0 },
|
|
{ "mli", dot_template, 0x2 }, /* old format, for compatibility */
|
|
{ "mlx", dot_template, 0x2 },
|
|
{ "mmi", dot_template, 0x4 },
|
|
{ "mfi", dot_template, 0x6 },
|
|
{ "mmf", dot_template, 0x7 },
|
|
{ "mib", dot_template, 0x8 },
|
|
{ "mbb", dot_template, 0x9 },
|
|
{ "bbb", dot_template, 0xb },
|
|
{ "mmb", dot_template, 0xc },
|
|
{ "mfb", dot_template, 0xe },
|
|
{ "align", dot_align, 0 },
|
|
{ "regstk", dot_regstk, 0 },
|
|
{ "rotr", dot_rot, DYNREG_GR },
|
|
{ "rotf", dot_rot, DYNREG_FR },
|
|
{ "rotp", dot_rot, DYNREG_PR },
|
|
{ "lsb", dot_byteorder, 0 },
|
|
{ "msb", dot_byteorder, 1 },
|
|
{ "psr", dot_psr, 0 },
|
|
{ "alias", dot_alias, 0 },
|
|
{ "secalias", dot_alias, 1 },
|
|
{ "ln", dot_ln, 0 }, /* source line info (for debugging) */
|
|
|
|
{ "xdata1", dot_xdata, 1 },
|
|
{ "xdata2", dot_xdata, 2 },
|
|
{ "xdata4", dot_xdata, 4 },
|
|
{ "xdata8", dot_xdata, 8 },
|
|
{ "xdata16", dot_xdata, 16 },
|
|
{ "xreal4", dot_xfloat_cons, 'f' },
|
|
{ "xreal8", dot_xfloat_cons, 'd' },
|
|
{ "xreal10", dot_xfloat_cons, 'x' },
|
|
{ "xreal16", dot_xfloat_cons, 'X' },
|
|
{ "xstring", dot_xstringer, 8 + 0 },
|
|
{ "xstringz", dot_xstringer, 8 + 1 },
|
|
|
|
/* unaligned versions: */
|
|
{ "xdata2.ua", dot_xdata_ua, 2 },
|
|
{ "xdata4.ua", dot_xdata_ua, 4 },
|
|
{ "xdata8.ua", dot_xdata_ua, 8 },
|
|
{ "xdata16.ua", dot_xdata_ua, 16 },
|
|
{ "xreal4.ua", dot_xfloat_cons_ua, 'f' },
|
|
{ "xreal8.ua", dot_xfloat_cons_ua, 'd' },
|
|
{ "xreal10.ua", dot_xfloat_cons_ua, 'x' },
|
|
{ "xreal16.ua", dot_xfloat_cons_ua, 'X' },
|
|
|
|
/* annotations/DV checking support */
|
|
{ "entry", dot_entry, 0 },
|
|
{ "mem.offset", dot_mem_offset, 0 },
|
|
{ "pred.rel", dot_pred_rel, 0 },
|
|
{ "pred.rel.clear", dot_pred_rel, 'c' },
|
|
{ "pred.rel.imply", dot_pred_rel, 'i' },
|
|
{ "pred.rel.mutex", dot_pred_rel, 'm' },
|
|
{ "pred.safe_across_calls", dot_pred_rel, 's' },
|
|
{ "reg.val", dot_reg_val, 0 },
|
|
{ "serialize.data", dot_serialize, 0 },
|
|
{ "serialize.instruction", dot_serialize, 1 },
|
|
{ "auto", dot_dv_mode, 'a' },
|
|
{ "explicit", dot_dv_mode, 'e' },
|
|
{ "default", dot_dv_mode, 'd' },
|
|
|
|
/* ??? These are needed to make gas/testsuite/gas/elf/ehopt.s work.
|
|
IA-64 aligns data allocation pseudo-ops by default, so we have to
|
|
tell it that these ones are supposed to be unaligned. Long term,
|
|
should rewrite so that only IA-64 specific data allocation pseudo-ops
|
|
are aligned by default. */
|
|
{"2byte", stmt_cons_ua, 2},
|
|
{"4byte", stmt_cons_ua, 4},
|
|
{"8byte", stmt_cons_ua, 8},
|
|
|
|
#ifdef TE_VMS
|
|
{"vms_common", obj_elf_vms_common, 0},
|
|
#endif
|
|
|
|
{ NULL, 0, 0 }
|
|
};
|
|
|
|
static const struct pseudo_opcode
|
|
{
|
|
const char *name;
|
|
void (*handler) (int);
|
|
int arg;
|
|
}
|
|
pseudo_opcode[] =
|
|
{
|
|
/* these are more like pseudo-ops, but don't start with a dot */
|
|
{ "data1", cons, 1 },
|
|
{ "data2", cons, 2 },
|
|
{ "data4", cons, 4 },
|
|
{ "data8", cons, 8 },
|
|
{ "data16", cons, 16 },
|
|
{ "real4", stmt_float_cons, 'f' },
|
|
{ "real8", stmt_float_cons, 'd' },
|
|
{ "real10", stmt_float_cons, 'x' },
|
|
{ "real16", stmt_float_cons, 'X' },
|
|
{ "string", stringer, 8 + 0 },
|
|
{ "stringz", stringer, 8 + 1 },
|
|
|
|
/* unaligned versions: */
|
|
{ "data2.ua", stmt_cons_ua, 2 },
|
|
{ "data4.ua", stmt_cons_ua, 4 },
|
|
{ "data8.ua", stmt_cons_ua, 8 },
|
|
{ "data16.ua", stmt_cons_ua, 16 },
|
|
{ "real4.ua", float_cons, 'f' },
|
|
{ "real8.ua", float_cons, 'd' },
|
|
{ "real10.ua", float_cons, 'x' },
|
|
{ "real16.ua", float_cons, 'X' },
|
|
};
|
|
|
|
/* Declare a register by creating a symbol for it and entering it in
|
|
the symbol table. */
|
|
|
|
static symbolS *
|
|
declare_register (const char *name, unsigned int regnum)
|
|
{
|
|
const char *err;
|
|
symbolS *sym;
|
|
|
|
sym = symbol_create (name, reg_section, regnum, &zero_address_frag);
|
|
|
|
err = hash_insert (md.reg_hash, S_GET_NAME (sym), (void *) sym);
|
|
if (err)
|
|
as_fatal ("Inserting \"%s\" into register table failed: %s",
|
|
name, err);
|
|
|
|
return sym;
|
|
}
|
|
|
|
static void
|
|
declare_register_set (const char *prefix,
|
|
unsigned int num_regs,
|
|
unsigned int base_regnum)
|
|
{
|
|
char name[8];
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < num_regs; ++i)
|
|
{
|
|
snprintf (name, sizeof (name), "%s%u", prefix, i);
|
|
declare_register (name, base_regnum + i);
|
|
}
|
|
}
|
|
|
|
static unsigned int
|
|
operand_width (enum ia64_opnd opnd)
|
|
{
|
|
const struct ia64_operand *odesc = &elf64_ia64_operands[opnd];
|
|
unsigned int bits = 0;
|
|
int i;
|
|
|
|
bits = 0;
|
|
for (i = 0; i < NELEMS (odesc->field) && odesc->field[i].bits; ++i)
|
|
bits += odesc->field[i].bits;
|
|
|
|
return bits;
|
|
}
|
|
|
|
static enum operand_match_result
|
|
operand_match (const struct ia64_opcode *idesc, int res_index, expressionS *e)
|
|
{
|
|
enum ia64_opnd opnd = idesc->operands[res_index];
|
|
int bits, relocatable = 0;
|
|
struct insn_fix *fix;
|
|
bfd_signed_vma val;
|
|
|
|
switch (opnd)
|
|
{
|
|
/* constants: */
|
|
|
|
case IA64_OPND_AR_CCV:
|
|
if (e->X_op == O_register && e->X_add_number == REG_AR + 32)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_AR_CSD:
|
|
if (e->X_op == O_register && e->X_add_number == REG_AR + 25)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_AR_PFS:
|
|
if (e->X_op == O_register && e->X_add_number == REG_AR + 64)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_GR0:
|
|
if (e->X_op == O_register && e->X_add_number == REG_GR + 0)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_IP:
|
|
if (e->X_op == O_register && e->X_add_number == REG_IP)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_PR:
|
|
if (e->X_op == O_register && e->X_add_number == REG_PR)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_PR_ROT:
|
|
if (e->X_op == O_register && e->X_add_number == REG_PR_ROT)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_PSR:
|
|
if (e->X_op == O_register && e->X_add_number == REG_PSR)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_PSR_L:
|
|
if (e->X_op == O_register && e->X_add_number == REG_PSR_L)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_PSR_UM:
|
|
if (e->X_op == O_register && e->X_add_number == REG_PSR_UM)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_C1:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if (e->X_add_number == 1)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_C8:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if (e->X_add_number == 8)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_C16:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if (e->X_add_number == 16)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
/* register operands: */
|
|
|
|
case IA64_OPND_AR3:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_AR
|
|
&& e->X_add_number < REG_AR + 128)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_B1:
|
|
case IA64_OPND_B2:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_BR
|
|
&& e->X_add_number < REG_BR + 8)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_CR3:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_CR
|
|
&& e->X_add_number < REG_CR + 128)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_DAHR3:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_DAHR
|
|
&& e->X_add_number < REG_DAHR + 8)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_F1:
|
|
case IA64_OPND_F2:
|
|
case IA64_OPND_F3:
|
|
case IA64_OPND_F4:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_FR
|
|
&& e->X_add_number < REG_FR + 128)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_P1:
|
|
case IA64_OPND_P2:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_P
|
|
&& e->X_add_number < REG_P + 64)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_R1:
|
|
case IA64_OPND_R2:
|
|
case IA64_OPND_R3:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_GR
|
|
&& e->X_add_number < REG_GR + 128)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_R3_2:
|
|
if (e->X_op == O_register && e->X_add_number >= REG_GR)
|
|
{
|
|
if (e->X_add_number < REG_GR + 4)
|
|
return OPERAND_MATCH;
|
|
else if (e->X_add_number < REG_GR + 128)
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
/* indirect operands: */
|
|
case IA64_OPND_CPUID_R3:
|
|
case IA64_OPND_DBR_R3:
|
|
case IA64_OPND_DTR_R3:
|
|
case IA64_OPND_ITR_R3:
|
|
case IA64_OPND_IBR_R3:
|
|
case IA64_OPND_MSR_R3:
|
|
case IA64_OPND_PKR_R3:
|
|
case IA64_OPND_PMC_R3:
|
|
case IA64_OPND_PMD_R3:
|
|
case IA64_OPND_DAHR_R3:
|
|
case IA64_OPND_RR_R3:
|
|
if (e->X_op == O_index && e->X_op_symbol
|
|
&& (S_GET_VALUE (e->X_op_symbol) - IND_CPUID
|
|
== opnd - IA64_OPND_CPUID_R3))
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_MR3:
|
|
if (e->X_op == O_index && !e->X_op_symbol)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
/* immediate operands: */
|
|
case IA64_OPND_CNT2a:
|
|
case IA64_OPND_LEN4:
|
|
case IA64_OPND_LEN6:
|
|
bits = operand_width (idesc->operands[res_index]);
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) (e->X_add_number - 1) < ((bfd_vma) 1 << bits))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_CNT2b:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) (e->X_add_number - 1) < 3)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_CNT2c:
|
|
val = e->X_add_number;
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((val == 0 || val == 7 || val == 15 || val == 16))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_SOR:
|
|
/* SOR must be an integer multiple of 8 */
|
|
if (e->X_op == O_constant && e->X_add_number & 0x7)
|
|
return OPERAND_OUT_OF_RANGE;
|
|
/* Fall through. */
|
|
case IA64_OPND_SOF:
|
|
case IA64_OPND_SOL:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) e->X_add_number <= 96)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMMU62:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) e->X_add_number < ((bfd_vma) 1 << 62))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
else
|
|
{
|
|
/* FIXME -- need 62-bit relocation type */
|
|
as_bad (_("62-bit relocation not yet implemented"));
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMMU64:
|
|
if (e->X_op == O_symbol || e->X_op == O_pseudo_fixup
|
|
|| e->X_op == O_subtract)
|
|
{
|
|
fix = CURR_SLOT.fixup + CURR_SLOT.num_fixups;
|
|
fix->code = BFD_RELOC_IA64_IMM64;
|
|
if (e->X_op != O_subtract)
|
|
{
|
|
fix->code = ia64_gen_real_reloc_type (e->X_op_symbol, fix->code);
|
|
if (e->X_op == O_pseudo_fixup)
|
|
e->X_op = O_symbol;
|
|
}
|
|
|
|
fix->opnd = idesc->operands[res_index];
|
|
fix->expr = *e;
|
|
fix->is_pcrel = 0;
|
|
++CURR_SLOT.num_fixups;
|
|
return OPERAND_MATCH;
|
|
}
|
|
else if (e->X_op == O_constant)
|
|
return OPERAND_MATCH;
|
|
break;
|
|
|
|
case IA64_OPND_IMMU5b:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
val = e->X_add_number;
|
|
if (val >= 32 && val <= 63)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_CCNT5:
|
|
case IA64_OPND_CNT5:
|
|
case IA64_OPND_CNT6:
|
|
case IA64_OPND_CPOS6a:
|
|
case IA64_OPND_CPOS6b:
|
|
case IA64_OPND_CPOS6c:
|
|
case IA64_OPND_IMMU2:
|
|
case IA64_OPND_IMMU7a:
|
|
case IA64_OPND_IMMU7b:
|
|
case IA64_OPND_IMMU16:
|
|
case IA64_OPND_IMMU19:
|
|
case IA64_OPND_IMMU21:
|
|
case IA64_OPND_IMMU24:
|
|
case IA64_OPND_MBTYPE4:
|
|
case IA64_OPND_MHTYPE8:
|
|
case IA64_OPND_POS6:
|
|
bits = operand_width (idesc->operands[res_index]);
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) e->X_add_number < ((bfd_vma) 1 << bits))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMMU9:
|
|
bits = operand_width (idesc->operands[res_index]);
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((bfd_vma) e->X_add_number < ((bfd_vma) 1 << bits))
|
|
{
|
|
int lobits = e->X_add_number & 0x3;
|
|
if (((bfd_vma) e->X_add_number & 0x3C) != 0 && lobits == 0)
|
|
e->X_add_number |= (bfd_vma) 0x3;
|
|
return OPERAND_MATCH;
|
|
}
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMM44:
|
|
/* least 16 bits must be zero */
|
|
if ((e->X_add_number & 0xffff) != 0)
|
|
/* XXX technically, this is wrong: we should not be issuing warning
|
|
messages until we're sure this instruction pattern is going to
|
|
be used! */
|
|
as_warn (_("lower 16 bits of mask ignored"));
|
|
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if (((e->X_add_number >= 0
|
|
&& (bfd_vma) e->X_add_number < ((bfd_vma) 1 << 44))
|
|
|| (e->X_add_number < 0
|
|
&& (bfd_vma) -e->X_add_number <= ((bfd_vma) 1 << 44))))
|
|
{
|
|
/* sign-extend */
|
|
if (e->X_add_number >= 0
|
|
&& (e->X_add_number & ((bfd_vma) 1 << 43)) != 0)
|
|
{
|
|
e->X_add_number |= ~(((bfd_vma) 1 << 44) - 1);
|
|
}
|
|
return OPERAND_MATCH;
|
|
}
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMM17:
|
|
/* bit 0 is a don't care (pr0 is hardwired to 1) */
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if (((e->X_add_number >= 0
|
|
&& (bfd_vma) e->X_add_number < ((bfd_vma) 1 << 17))
|
|
|| (e->X_add_number < 0
|
|
&& (bfd_vma) -e->X_add_number <= ((bfd_vma) 1 << 17))))
|
|
{
|
|
/* sign-extend */
|
|
if (e->X_add_number >= 0
|
|
&& (e->X_add_number & ((bfd_vma) 1 << 16)) != 0)
|
|
{
|
|
e->X_add_number |= ~(((bfd_vma) 1 << 17) - 1);
|
|
}
|
|
return OPERAND_MATCH;
|
|
}
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_IMM14:
|
|
case IA64_OPND_IMM22:
|
|
relocatable = 1;
|
|
/* Fall through. */
|
|
case IA64_OPND_IMM1:
|
|
case IA64_OPND_IMM8:
|
|
case IA64_OPND_IMM8U4:
|
|
case IA64_OPND_IMM8M1:
|
|
case IA64_OPND_IMM8M1U4:
|
|
case IA64_OPND_IMM8M1U8:
|
|
case IA64_OPND_IMM9a:
|
|
case IA64_OPND_IMM9b:
|
|
bits = operand_width (idesc->operands[res_index]);
|
|
if (relocatable && (e->X_op == O_symbol
|
|
|| e->X_op == O_subtract
|
|
|| e->X_op == O_pseudo_fixup))
|
|
{
|
|
fix = CURR_SLOT.fixup + CURR_SLOT.num_fixups;
|
|
|
|
if (idesc->operands[res_index] == IA64_OPND_IMM14)
|
|
fix->code = BFD_RELOC_IA64_IMM14;
|
|
else
|
|
fix->code = BFD_RELOC_IA64_IMM22;
|
|
|
|
if (e->X_op != O_subtract)
|
|
{
|
|
fix->code = ia64_gen_real_reloc_type (e->X_op_symbol, fix->code);
|
|
if (e->X_op == O_pseudo_fixup)
|
|
e->X_op = O_symbol;
|
|
}
|
|
|
|
fix->opnd = idesc->operands[res_index];
|
|
fix->expr = *e;
|
|
fix->is_pcrel = 0;
|
|
++CURR_SLOT.num_fixups;
|
|
return OPERAND_MATCH;
|
|
}
|
|
else if (e->X_op != O_constant
|
|
&& ! (e->X_op == O_big && opnd == IA64_OPND_IMM8M1U8))
|
|
return OPERAND_MISMATCH;
|
|
|
|
if (opnd == IA64_OPND_IMM8M1U4)
|
|
{
|
|
/* Zero is not valid for unsigned compares that take an adjusted
|
|
constant immediate range. */
|
|
if (e->X_add_number == 0)
|
|
return OPERAND_OUT_OF_RANGE;
|
|
|
|
/* Sign-extend 32-bit unsigned numbers, so that the following range
|
|
checks will work. */
|
|
val = e->X_add_number;
|
|
if (((val & (~(bfd_vma) 0 << 32)) == 0)
|
|
&& ((val & ((bfd_vma) 1 << 31)) != 0))
|
|
val = ((val << 32) >> 32);
|
|
|
|
/* Check for 0x100000000. This is valid because
|
|
0x100000000-1 is the same as ((uint32_t) -1). */
|
|
if (val == ((bfd_signed_vma) 1 << 32))
|
|
return OPERAND_MATCH;
|
|
|
|
val = val - 1;
|
|
}
|
|
else if (opnd == IA64_OPND_IMM8M1U8)
|
|
{
|
|
/* Zero is not valid for unsigned compares that take an adjusted
|
|
constant immediate range. */
|
|
if (e->X_add_number == 0)
|
|
return OPERAND_OUT_OF_RANGE;
|
|
|
|
/* Check for 0x10000000000000000. */
|
|
if (e->X_op == O_big)
|
|
{
|
|
if (generic_bignum[0] == 0
|
|
&& generic_bignum[1] == 0
|
|
&& generic_bignum[2] == 0
|
|
&& generic_bignum[3] == 0
|
|
&& generic_bignum[4] == 1)
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
else
|
|
val = e->X_add_number - 1;
|
|
}
|
|
else if (opnd == IA64_OPND_IMM8M1)
|
|
val = e->X_add_number - 1;
|
|
else if (opnd == IA64_OPND_IMM8U4)
|
|
{
|
|
/* Sign-extend 32-bit unsigned numbers, so that the following range
|
|
checks will work. */
|
|
val = e->X_add_number;
|
|
if (((val & (~(bfd_vma) 0 << 32)) == 0)
|
|
&& ((val & ((bfd_vma) 1 << 31)) != 0))
|
|
val = ((val << 32) >> 32);
|
|
}
|
|
else
|
|
val = e->X_add_number;
|
|
|
|
if ((val >= 0 && (bfd_vma) val < ((bfd_vma) 1 << (bits - 1)))
|
|
|| (val < 0 && (bfd_vma) -val <= ((bfd_vma) 1 << (bits - 1))))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
|
|
case IA64_OPND_INC3:
|
|
/* +/- 1, 4, 8, 16 */
|
|
val = e->X_add_number;
|
|
if (val < 0)
|
|
val = -val;
|
|
if (e->X_op == O_constant)
|
|
{
|
|
if ((val == 1 || val == 4 || val == 8 || val == 16))
|
|
return OPERAND_MATCH;
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_TGT25:
|
|
case IA64_OPND_TGT25b:
|
|
case IA64_OPND_TGT25c:
|
|
case IA64_OPND_TGT64:
|
|
if (e->X_op == O_symbol)
|
|
{
|
|
fix = CURR_SLOT.fixup + CURR_SLOT.num_fixups;
|
|
if (opnd == IA64_OPND_TGT25)
|
|
fix->code = BFD_RELOC_IA64_PCREL21F;
|
|
else if (opnd == IA64_OPND_TGT25b)
|
|
fix->code = BFD_RELOC_IA64_PCREL21M;
|
|
else if (opnd == IA64_OPND_TGT25c)
|
|
fix->code = BFD_RELOC_IA64_PCREL21B;
|
|
else if (opnd == IA64_OPND_TGT64)
|
|
fix->code = BFD_RELOC_IA64_PCREL60B;
|
|
else
|
|
abort ();
|
|
|
|
fix->code = ia64_gen_real_reloc_type (e->X_op_symbol, fix->code);
|
|
fix->opnd = idesc->operands[res_index];
|
|
fix->expr = *e;
|
|
fix->is_pcrel = 1;
|
|
++CURR_SLOT.num_fixups;
|
|
return OPERAND_MATCH;
|
|
}
|
|
/* Fall through. */
|
|
case IA64_OPND_TAG13:
|
|
case IA64_OPND_TAG13b:
|
|
switch (e->X_op)
|
|
{
|
|
case O_constant:
|
|
return OPERAND_MATCH;
|
|
|
|
case O_symbol:
|
|
fix = CURR_SLOT.fixup + CURR_SLOT.num_fixups;
|
|
/* There are no external relocs for TAG13/TAG13b fields, so we
|
|
create a dummy reloc. This will not live past md_apply_fix. */
|
|
fix->code = BFD_RELOC_UNUSED;
|
|
fix->code = ia64_gen_real_reloc_type (e->X_op_symbol, fix->code);
|
|
fix->opnd = idesc->operands[res_index];
|
|
fix->expr = *e;
|
|
fix->is_pcrel = 1;
|
|
++CURR_SLOT.num_fixups;
|
|
return OPERAND_MATCH;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
break;
|
|
|
|
case IA64_OPND_LDXMOV:
|
|
fix = CURR_SLOT.fixup + CURR_SLOT.num_fixups;
|
|
fix->code = BFD_RELOC_IA64_LDXMOV;
|
|
fix->opnd = idesc->operands[res_index];
|
|
fix->expr = *e;
|
|
fix->is_pcrel = 0;
|
|
++CURR_SLOT.num_fixups;
|
|
return OPERAND_MATCH;
|
|
|
|
case IA64_OPND_STRD5b:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
/* 5-bit signed scaled by 64 */
|
|
if ((e->X_add_number <= ( 0xf << 6 ))
|
|
&& (e->X_add_number >= -( 0x10 << 6 )))
|
|
{
|
|
|
|
/* Must be a multiple of 64 */
|
|
if ((e->X_add_number & 0x3f) != 0)
|
|
as_warn (_("stride must be a multiple of 64; lower 6 bits ignored"));
|
|
|
|
e->X_add_number &= ~ 0x3f;
|
|
return OPERAND_MATCH;
|
|
}
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
case IA64_OPND_CNT6a:
|
|
if (e->X_op == O_constant)
|
|
{
|
|
/* 6-bit unsigned biased by 1 -- count 0 is meaningless */
|
|
if ((e->X_add_number <= 64)
|
|
&& (e->X_add_number > 0) )
|
|
{
|
|
return OPERAND_MATCH;
|
|
}
|
|
else
|
|
return OPERAND_OUT_OF_RANGE;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
return OPERAND_MISMATCH;
|
|
}
|
|
|
|
static int
|
|
parse_operand (expressionS *e, int more)
|
|
{
|
|
int sep = '\0';
|
|
|
|
memset (e, 0, sizeof (*e));
|
|
e->X_op = O_absent;
|
|
SKIP_WHITESPACE ();
|
|
expression (e);
|
|
sep = *input_line_pointer;
|
|
if (more && (sep == ',' || sep == more))
|
|
++input_line_pointer;
|
|
return sep;
|
|
}
|
|
|
|
static int
|
|
parse_operand_and_eval (expressionS *e, int more)
|
|
{
|
|
int sep = parse_operand (e, more);
|
|
resolve_expression (e);
|
|
return sep;
|
|
}
|
|
|
|
static int
|
|
parse_operand_maybe_eval (expressionS *e, int more, enum ia64_opnd op)
|
|
{
|
|
int sep = parse_operand (e, more);
|
|
switch (op)
|
|
{
|
|
case IA64_OPND_IMM14:
|
|
case IA64_OPND_IMM22:
|
|
case IA64_OPND_IMMU64:
|
|
case IA64_OPND_TGT25:
|
|
case IA64_OPND_TGT25b:
|
|
case IA64_OPND_TGT25c:
|
|
case IA64_OPND_TGT64:
|
|
case IA64_OPND_TAG13:
|
|
case IA64_OPND_TAG13b:
|
|
case IA64_OPND_LDXMOV:
|
|
break;
|
|
default:
|
|
resolve_expression (e);
|
|
break;
|
|
}
|
|
return sep;
|
|
}
|
|
|
|
/* Returns the next entry in the opcode table that matches the one in
|
|
IDESC, and frees the entry in IDESC. If no matching entry is
|
|
found, NULL is returned instead. */
|
|
|
|
static struct ia64_opcode *
|
|
get_next_opcode (struct ia64_opcode *idesc)
|
|
{
|
|
struct ia64_opcode *next = ia64_find_next_opcode (idesc);
|
|
ia64_free_opcode (idesc);
|
|
return next;
|
|
}
|
|
|
|
/* Parse the operands for the opcode and find the opcode variant that
|
|
matches the specified operands, or NULL if no match is possible. */
|
|
|
|
static struct ia64_opcode *
|
|
parse_operands (struct ia64_opcode *idesc)
|
|
{
|
|
int i = 0, highest_unmatched_operand, num_operands = 0, num_outputs = 0;
|
|
int error_pos, out_of_range_pos, curr_out_of_range_pos, sep = 0;
|
|
int reg1, reg2;
|
|
char reg_class;
|
|
enum ia64_opnd expected_operand = IA64_OPND_NIL;
|
|
enum operand_match_result result;
|
|
char mnemonic[129];
|
|
char *first_arg = 0, *end, *saved_input_pointer;
|
|
unsigned int sof;
|
|
|
|
gas_assert (strlen (idesc->name) <= 128);
|
|
|
|
strcpy (mnemonic, idesc->name);
|
|
if (idesc->operands[2] == IA64_OPND_SOF
|
|
|| idesc->operands[1] == IA64_OPND_SOF)
|
|
{
|
|
/* To make the common idiom "alloc loc?=ar.pfs,0,1,0,0" work, we
|
|
can't parse the first operand until we have parsed the
|
|
remaining operands of the "alloc" instruction. */
|
|
SKIP_WHITESPACE ();
|
|
first_arg = input_line_pointer;
|
|
end = strchr (input_line_pointer, '=');
|
|
if (!end)
|
|
{
|
|
as_bad (_("Expected separator `='"));
|
|
return 0;
|
|
}
|
|
input_line_pointer = end + 1;
|
|
++i;
|
|
++num_outputs;
|
|
}
|
|
|
|
for (; ; ++i)
|
|
{
|
|
if (i < NELEMS (CURR_SLOT.opnd))
|
|
{
|
|
sep = parse_operand_maybe_eval (CURR_SLOT.opnd + i, '=',
|
|
idesc->operands[i]);
|
|
if (CURR_SLOT.opnd[i].X_op == O_absent)
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
expressionS dummy;
|
|
|
|
sep = parse_operand (&dummy, '=');
|
|
if (dummy.X_op == O_absent)
|
|
break;
|
|
}
|
|
|
|
++num_operands;
|
|
|
|
if (sep != '=' && sep != ',')
|
|
break;
|
|
|
|
if (sep == '=')
|
|
{
|
|
if (num_outputs > 0)
|
|
as_bad (_("Duplicate equal sign (=) in instruction"));
|
|
else
|
|
num_outputs = i + 1;
|
|
}
|
|
}
|
|
if (sep != '\0')
|
|
{
|
|
as_bad (_("Illegal operand separator `%c'"), sep);
|
|
return 0;
|
|
}
|
|
|
|
if (idesc->operands[2] == IA64_OPND_SOF
|
|
|| idesc->operands[1] == IA64_OPND_SOF)
|
|
{
|
|
/* Map alloc r1=ar.pfs,i,l,o,r to alloc r1=ar.pfs,(i+l+o),(i+l),r.
|
|
Note, however, that due to that mapping operand numbers in error
|
|
messages for any of the constant operands will not be correct. */
|
|
know (strcmp (idesc->name, "alloc") == 0);
|
|
/* The first operand hasn't been parsed/initialized, yet (but
|
|
num_operands intentionally doesn't account for that). */
|
|
i = num_operands > 4 ? 2 : 1;
|
|
#define FORCE_CONST(n) (CURR_SLOT.opnd[n].X_op == O_constant \
|
|
? CURR_SLOT.opnd[n].X_add_number \
|
|
: 0)
|
|
sof = set_regstack (FORCE_CONST(i),
|
|
FORCE_CONST(i + 1),
|
|
FORCE_CONST(i + 2),
|
|
FORCE_CONST(i + 3));
|
|
#undef FORCE_CONST
|
|
|
|
/* now we can parse the first arg: */
|
|
saved_input_pointer = input_line_pointer;
|
|
input_line_pointer = first_arg;
|
|
sep = parse_operand_maybe_eval (CURR_SLOT.opnd + 0, '=',
|
|
idesc->operands[0]);
|
|
if (sep != '=')
|
|
--num_outputs; /* force error */
|
|
input_line_pointer = saved_input_pointer;
|
|
|
|
CURR_SLOT.opnd[i].X_add_number = sof;
|
|
if (CURR_SLOT.opnd[i + 1].X_op == O_constant
|
|
&& CURR_SLOT.opnd[i + 2].X_op == O_constant)
|
|
CURR_SLOT.opnd[i + 1].X_add_number
|
|
= sof - CURR_SLOT.opnd[i + 2].X_add_number;
|
|
else
|
|
CURR_SLOT.opnd[i + 1].X_op = O_illegal;
|
|
CURR_SLOT.opnd[i + 2] = CURR_SLOT.opnd[i + 3];
|
|
}
|
|
|
|
highest_unmatched_operand = -4;
|
|
curr_out_of_range_pos = -1;
|
|
error_pos = 0;
|
|
for (; idesc; idesc = get_next_opcode (idesc))
|
|
{
|
|
if (num_outputs != idesc->num_outputs)
|
|
continue; /* mismatch in # of outputs */
|
|
if (highest_unmatched_operand < 0)
|
|
highest_unmatched_operand |= 1;
|
|
if (num_operands > NELEMS (idesc->operands)
|
|
|| (num_operands < NELEMS (idesc->operands)
|
|
&& idesc->operands[num_operands])
|
|
|| (num_operands > 0 && !idesc->operands[num_operands - 1]))
|
|
continue; /* mismatch in number of arguments */
|
|
if (highest_unmatched_operand < 0)
|
|
highest_unmatched_operand |= 2;
|
|
|
|
CURR_SLOT.num_fixups = 0;
|
|
|
|
/* Try to match all operands. If we see an out-of-range operand,
|
|
then continue trying to match the rest of the operands, since if
|
|
the rest match, then this idesc will give the best error message. */
|
|
|
|
out_of_range_pos = -1;
|
|
for (i = 0; i < num_operands && idesc->operands[i]; ++i)
|
|
{
|
|
result = operand_match (idesc, i, CURR_SLOT.opnd + i);
|
|
if (result != OPERAND_MATCH)
|
|
{
|
|
if (result != OPERAND_OUT_OF_RANGE)
|
|
break;
|
|
if (out_of_range_pos < 0)
|
|
/* remember position of the first out-of-range operand: */
|
|
out_of_range_pos = i;
|
|
}
|
|
}
|
|
|
|
/* If we did not match all operands, or if at least one operand was
|
|
out-of-range, then this idesc does not match. Keep track of which
|
|
idesc matched the most operands before failing. If we have two
|
|
idescs that failed at the same position, and one had an out-of-range
|
|
operand, then prefer the out-of-range operand. Thus if we have
|
|
"add r0=0x1000000,r1" we get an error saying the constant is out
|
|
of range instead of an error saying that the constant should have been
|
|
a register. */
|
|
|
|
if (i != num_operands || out_of_range_pos >= 0)
|
|
{
|
|
if (i > highest_unmatched_operand
|
|
|| (i == highest_unmatched_operand
|
|
&& out_of_range_pos > curr_out_of_range_pos))
|
|
{
|
|
highest_unmatched_operand = i;
|
|
if (out_of_range_pos >= 0)
|
|
{
|
|
expected_operand = idesc->operands[out_of_range_pos];
|
|
error_pos = out_of_range_pos;
|
|
}
|
|
else
|
|
{
|
|
expected_operand = idesc->operands[i];
|
|
error_pos = i;
|
|
}
|
|
curr_out_of_range_pos = out_of_range_pos;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
break;
|
|
}
|
|
if (!idesc)
|
|
{
|
|
if (expected_operand)
|
|
as_bad (_("Operand %u of `%s' should be %s"),
|
|
error_pos + 1, mnemonic,
|
|
elf64_ia64_operands[expected_operand].desc);
|
|
else if (highest_unmatched_operand < 0 && !(highest_unmatched_operand & 1))
|
|
as_bad (_("Wrong number of output operands"));
|
|
else if (highest_unmatched_operand < 0 && !(highest_unmatched_operand & 2))
|
|
as_bad (_("Wrong number of input operands"));
|
|
else
|
|
as_bad (_("Operand mismatch"));
|
|
return 0;
|
|
}
|
|
|
|
/* Check that the instruction doesn't use
|
|
- r0, f0, or f1 as output operands
|
|
- the same predicate twice as output operands
|
|
- r0 as address of a base update load or store
|
|
- the same GR as output and address of a base update load
|
|
- two even- or two odd-numbered FRs as output operands of a floating
|
|
point parallel load.
|
|
At most two (conflicting) output (or output-like) operands can exist,
|
|
(floating point parallel loads have three outputs, but the base register,
|
|
if updated, cannot conflict with the actual outputs). */
|
|
reg2 = reg1 = -1;
|
|
for (i = 0; i < num_operands; ++i)
|
|
{
|
|
int regno = 0;
|
|
|
|
reg_class = 0;
|
|
switch (idesc->operands[i])
|
|
{
|
|
case IA64_OPND_R1:
|
|
case IA64_OPND_R2:
|
|
case IA64_OPND_R3:
|
|
if (i < num_outputs)
|
|
{
|
|
if (CURR_SLOT.opnd[i].X_add_number == REG_GR)
|
|
reg_class = 'r';
|
|
else if (reg1 < 0)
|
|
reg1 = CURR_SLOT.opnd[i].X_add_number;
|
|
else if (reg2 < 0)
|
|
reg2 = CURR_SLOT.opnd[i].X_add_number;
|
|
}
|
|
break;
|
|
case IA64_OPND_P1:
|
|
case IA64_OPND_P2:
|
|
if (i < num_outputs)
|
|
{
|
|
if (reg1 < 0)
|
|
reg1 = CURR_SLOT.opnd[i].X_add_number;
|
|
else if (reg2 < 0)
|
|
reg2 = CURR_SLOT.opnd[i].X_add_number;
|
|
}
|
|
break;
|
|
case IA64_OPND_F1:
|
|
case IA64_OPND_F2:
|
|
case IA64_OPND_F3:
|
|
case IA64_OPND_F4:
|
|
if (i < num_outputs)
|
|
{
|
|
if (CURR_SLOT.opnd[i].X_add_number >= REG_FR
|
|
&& CURR_SLOT.opnd[i].X_add_number <= REG_FR + 1)
|
|
{
|
|
reg_class = 'f';
|
|
regno = CURR_SLOT.opnd[i].X_add_number - REG_FR;
|
|
}
|
|
else if (reg1 < 0)
|
|
reg1 = CURR_SLOT.opnd[i].X_add_number;
|
|
else if (reg2 < 0)
|
|
reg2 = CURR_SLOT.opnd[i].X_add_number;
|
|
}
|
|
break;
|
|
case IA64_OPND_MR3:
|
|
if (idesc->flags & IA64_OPCODE_POSTINC)
|
|
{
|
|
if (CURR_SLOT.opnd[i].X_add_number == REG_GR)
|
|
reg_class = 'm';
|
|
else if (reg1 < 0)
|
|
reg1 = CURR_SLOT.opnd[i].X_add_number;
|
|
else if (reg2 < 0)
|
|
reg2 = CURR_SLOT.opnd[i].X_add_number;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
switch (reg_class)
|
|
{
|
|
case 0:
|
|
break;
|
|
default:
|
|
as_warn (_("Invalid use of `%c%d' as output operand"), reg_class, regno);
|
|
break;
|
|
case 'm':
|
|
as_warn (_("Invalid use of `r%d' as base update address operand"), regno);
|
|
break;
|
|
}
|
|
}
|
|
if (reg1 == reg2)
|
|
{
|
|
if (reg1 >= REG_GR && reg1 <= REG_GR + 127)
|
|
{
|
|
reg1 -= REG_GR;
|
|
reg_class = 'r';
|
|
}
|
|
else if (reg1 >= REG_P && reg1 <= REG_P + 63)
|
|
{
|
|
reg1 -= REG_P;
|
|
reg_class = 'p';
|
|
}
|
|
else if (reg1 >= REG_FR && reg1 <= REG_FR + 127)
|
|
{
|
|
reg1 -= REG_FR;
|
|
reg_class = 'f';
|
|
}
|
|
else
|
|
reg_class = 0;
|
|
if (reg_class)
|
|
as_warn (_("Invalid duplicate use of `%c%d'"), reg_class, reg1);
|
|
}
|
|
else if (((reg1 >= REG_FR && reg1 <= REG_FR + 31
|
|
&& reg2 >= REG_FR && reg2 <= REG_FR + 31)
|
|
|| (reg1 >= REG_FR + 32 && reg1 <= REG_FR + 127
|
|
&& reg2 >= REG_FR + 32 && reg2 <= REG_FR + 127))
|
|
&& ! ((reg1 ^ reg2) & 1))
|
|
as_warn (_("Invalid simultaneous use of `f%d' and `f%d'"),
|
|
reg1 - REG_FR, reg2 - REG_FR);
|
|
else if ((reg1 >= REG_FR && reg1 <= REG_FR + 31
|
|
&& reg2 >= REG_FR + 32 && reg2 <= REG_FR + 127)
|
|
|| (reg1 >= REG_FR + 32 && reg1 <= REG_FR + 127
|
|
&& reg2 >= REG_FR && reg2 <= REG_FR + 31))
|
|
as_warn (_("Dangerous simultaneous use of `f%d' and `f%d'"),
|
|
reg1 - REG_FR, reg2 - REG_FR);
|
|
return idesc;
|
|
}
|
|
|
|
static void
|
|
build_insn (struct slot *slot, bfd_vma *insnp)
|
|
{
|
|
const struct ia64_operand *odesc, *o2desc;
|
|
struct ia64_opcode *idesc = slot->idesc;
|
|
bfd_vma insn;
|
|
bfd_signed_vma val;
|
|
const char *err;
|
|
int i;
|
|
|
|
insn = idesc->opcode | slot->qp_regno;
|
|
|
|
for (i = 0; i < NELEMS (idesc->operands) && idesc->operands[i]; ++i)
|
|
{
|
|
if (slot->opnd[i].X_op == O_register
|
|
|| slot->opnd[i].X_op == O_constant
|
|
|| slot->opnd[i].X_op == O_index)
|
|
val = slot->opnd[i].X_add_number;
|
|
else if (slot->opnd[i].X_op == O_big)
|
|
{
|
|
/* This must be the value 0x10000000000000000. */
|
|
gas_assert (idesc->operands[i] == IA64_OPND_IMM8M1U8);
|
|
val = 0;
|
|
}
|
|
else
|
|
val = 0;
|
|
|
|
switch (idesc->operands[i])
|
|
{
|
|
case IA64_OPND_IMMU64:
|
|
*insnp++ = (val >> 22) & 0x1ffffffffffLL;
|
|
insn |= (((val & 0x7f) << 13) | (((val >> 7) & 0x1ff) << 27)
|
|
| (((val >> 16) & 0x1f) << 22) | (((val >> 21) & 0x1) << 21)
|
|
| (((val >> 63) & 0x1) << 36));
|
|
continue;
|
|
|
|
case IA64_OPND_IMMU62:
|
|
val &= 0x3fffffffffffffffULL;
|
|
if (val != slot->opnd[i].X_add_number)
|
|
as_warn (_("Value truncated to 62 bits"));
|
|
*insnp++ = (val >> 21) & 0x1ffffffffffLL;
|
|
insn |= (((val & 0xfffff) << 6) | (((val >> 20) & 0x1) << 36));
|
|
continue;
|
|
|
|
case IA64_OPND_TGT64:
|
|
val >>= 4;
|
|
*insnp++ = ((val >> 20) & 0x7fffffffffLL) << 2;
|
|
insn |= ((((val >> 59) & 0x1) << 36)
|
|
| (((val >> 0) & 0xfffff) << 13));
|
|
continue;
|
|
|
|
case IA64_OPND_AR3:
|
|
val -= REG_AR;
|
|
break;
|
|
|
|
case IA64_OPND_B1:
|
|
case IA64_OPND_B2:
|
|
val -= REG_BR;
|
|
break;
|
|
|
|
case IA64_OPND_CR3:
|
|
val -= REG_CR;
|
|
break;
|
|
|
|
case IA64_OPND_DAHR3:
|
|
val -= REG_DAHR;
|
|
break;
|
|
|
|
case IA64_OPND_F1:
|
|
case IA64_OPND_F2:
|
|
case IA64_OPND_F3:
|
|
case IA64_OPND_F4:
|
|
val -= REG_FR;
|
|
break;
|
|
|
|
case IA64_OPND_P1:
|
|
case IA64_OPND_P2:
|
|
val -= REG_P;
|
|
break;
|
|
|
|
case IA64_OPND_R1:
|
|
case IA64_OPND_R2:
|
|
case IA64_OPND_R3:
|
|
case IA64_OPND_R3_2:
|
|
case IA64_OPND_CPUID_R3:
|
|
case IA64_OPND_DBR_R3:
|
|
case IA64_OPND_DTR_R3:
|
|
case IA64_OPND_ITR_R3:
|
|
case IA64_OPND_IBR_R3:
|
|
case IA64_OPND_MR3:
|
|
case IA64_OPND_MSR_R3:
|
|
case IA64_OPND_PKR_R3:
|
|
case IA64_OPND_PMC_R3:
|
|
case IA64_OPND_PMD_R3:
|
|
case IA64_OPND_DAHR_R3:
|
|
case IA64_OPND_RR_R3:
|
|
val -= REG_GR;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
odesc = elf64_ia64_operands + idesc->operands[i];
|
|
err = (*odesc->insert) (odesc, val, &insn);
|
|
if (err)
|
|
as_bad_where (slot->src_file, slot->src_line,
|
|
_("Bad operand value: %s"), err);
|
|
if (idesc->flags & IA64_OPCODE_PSEUDO)
|
|
{
|
|
if ((idesc->flags & IA64_OPCODE_F2_EQ_F3)
|
|
&& odesc == elf64_ia64_operands + IA64_OPND_F3)
|
|
{
|
|
o2desc = elf64_ia64_operands + IA64_OPND_F2;
|
|
(*o2desc->insert) (o2desc, val, &insn);
|
|
}
|
|
if ((idesc->flags & IA64_OPCODE_LEN_EQ_64MCNT)
|
|
&& (odesc == elf64_ia64_operands + IA64_OPND_CPOS6a
|
|
|| odesc == elf64_ia64_operands + IA64_OPND_POS6))
|
|
{
|
|
o2desc = elf64_ia64_operands + IA64_OPND_LEN6;
|
|
(*o2desc->insert) (o2desc, 64 - val, &insn);
|
|
}
|
|
}
|
|
}
|
|
*insnp = insn;
|
|
}
|
|
|
|
static void
|
|
emit_one_bundle (void)
|
|
{
|
|
int manual_bundling_off = 0, manual_bundling = 0;
|
|
enum ia64_unit required_unit, insn_unit = 0;
|
|
enum ia64_insn_type type[3], insn_type;
|
|
unsigned int template_val, orig_template;
|
|
bfd_vma insn[3] = { -1, -1, -1 };
|
|
struct ia64_opcode *idesc;
|
|
int end_of_insn_group = 0, user_template = -1;
|
|
int n, i, j, first, curr, last_slot;
|
|
bfd_vma t0 = 0, t1 = 0;
|
|
struct label_fix *lfix;
|
|
bfd_boolean mark_label;
|
|
struct insn_fix *ifix;
|
|
char mnemonic[16];
|
|
fixS *fix;
|
|
char *f;
|
|
int addr_mod;
|
|
|
|
first = (md.curr_slot + NUM_SLOTS - md.num_slots_in_use) % NUM_SLOTS;
|
|
know (first >= 0 && first < NUM_SLOTS);
|
|
n = MIN (3, md.num_slots_in_use);
|
|
|
|
/* Determine template: user user_template if specified, best match
|
|
otherwise: */
|
|
|
|
if (md.slot[first].user_template >= 0)
|
|
user_template = template_val = md.slot[first].user_template;
|
|
else
|
|
{
|
|
/* Auto select appropriate template. */
|
|
memset (type, 0, sizeof (type));
|
|
curr = first;
|
|
for (i = 0; i < n; ++i)
|
|
{
|
|
if (md.slot[curr].label_fixups && i != 0)
|
|
break;
|
|
type[i] = md.slot[curr].idesc->type;
|
|
curr = (curr + 1) % NUM_SLOTS;
|
|
}
|
|
template_val = best_template[type[0]][type[1]][type[2]];
|
|
}
|
|
|
|
/* initialize instructions with appropriate nops: */
|
|
for (i = 0; i < 3; ++i)
|
|
insn[i] = nop[ia64_templ_desc[template_val].exec_unit[i]];
|
|
|
|
f = frag_more (16);
|
|
|
|
/* Check to see if this bundle is at an offset that is a multiple of 16-bytes
|
|
from the start of the frag. */
|
|
addr_mod = frag_now_fix () & 15;
|
|
if (frag_now->has_code && frag_now->insn_addr != addr_mod)
|
|
as_bad (_("instruction address is not a multiple of 16"));
|
|
frag_now->insn_addr = addr_mod;
|
|
frag_now->has_code = 1;
|
|
|
|
/* now fill in slots with as many insns as possible: */
|
|
curr = first;
|
|
idesc = md.slot[curr].idesc;
|
|
end_of_insn_group = 0;
|
|
last_slot = -1;
|
|
for (i = 0; i < 3 && md.num_slots_in_use > 0; ++i)
|
|
{
|
|
/* If we have unwind records, we may need to update some now. */
|
|
unw_rec_list *ptr = md.slot[curr].unwind_record;
|
|
unw_rec_list *end_ptr = NULL;
|
|
|
|
if (ptr)
|
|
{
|
|
/* Find the last prologue/body record in the list for the current
|
|
insn, and set the slot number for all records up to that point.
|
|
This needs to be done now, because prologue/body records refer to
|
|
the current point, not the point after the instruction has been
|
|
issued. This matters because there may have been nops emitted
|
|
meanwhile. Any non-prologue non-body record followed by a
|
|
prologue/body record must also refer to the current point. */
|
|
unw_rec_list *last_ptr;
|
|
|
|
for (j = 1; end_ptr == NULL && j < md.num_slots_in_use; ++j)
|
|
end_ptr = md.slot[(curr + j) % NUM_SLOTS].unwind_record;
|
|
for (last_ptr = NULL; ptr != end_ptr; ptr = ptr->next)
|
|
if (ptr->r.type == prologue || ptr->r.type == prologue_gr
|
|
|| ptr->r.type == body)
|
|
last_ptr = ptr;
|
|
if (last_ptr)
|
|
{
|
|
/* Make last_ptr point one after the last prologue/body
|
|
record. */
|
|
last_ptr = last_ptr->next;
|
|
for (ptr = md.slot[curr].unwind_record; ptr != last_ptr;
|
|
ptr = ptr->next)
|
|
{
|
|
ptr->slot_number = (unsigned long) f + i;
|
|
ptr->slot_frag = frag_now;
|
|
}
|
|
/* Remove the initialized records, so that we won't accidentally
|
|
update them again if we insert a nop and continue. */
|
|
md.slot[curr].unwind_record = last_ptr;
|
|
}
|
|
}
|
|
|
|
manual_bundling_off = md.slot[curr].manual_bundling_off;
|
|
if (md.slot[curr].manual_bundling_on)
|
|
{
|
|
if (curr == first)
|
|
manual_bundling = 1;
|
|
else
|
|
break; /* Need to start a new bundle. */
|
|
}
|
|
|
|
/* If this instruction specifies a template, then it must be the first
|
|
instruction of a bundle. */
|
|
if (curr != first && md.slot[curr].user_template >= 0)
|
|
break;
|
|
|
|
if (idesc->flags & IA64_OPCODE_SLOT2)
|
|
{
|
|
if (manual_bundling && !manual_bundling_off)
|
|
{
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("`%s' must be last in bundle"), idesc->name);
|
|
if (i < 2)
|
|
manual_bundling = -1; /* Suppress meaningless post-loop errors. */
|
|
}
|
|
i = 2;
|
|
}
|
|
if (idesc->flags & IA64_OPCODE_LAST)
|
|
{
|
|
int required_slot;
|
|
unsigned int required_template;
|
|
|
|
/* If we need a stop bit after an M slot, our only choice is
|
|
template 5 (M;;MI). If we need a stop bit after a B
|
|
slot, our only choice is to place it at the end of the
|
|
bundle, because the only available templates are MIB,
|
|
MBB, BBB, MMB, and MFB. We don't handle anything other
|
|
than M and B slots because these are the only kind of
|
|
instructions that can have the IA64_OPCODE_LAST bit set. */
|
|
required_template = template_val;
|
|
switch (idesc->type)
|
|
{
|
|
case IA64_TYPE_M:
|
|
required_slot = 0;
|
|
required_template = 5;
|
|
break;
|
|
|
|
case IA64_TYPE_B:
|
|
required_slot = 2;
|
|
break;
|
|
|
|
default:
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("Internal error: don't know how to force %s to end of instruction group"),
|
|
idesc->name);
|
|
required_slot = i;
|
|
break;
|
|
}
|
|
if (manual_bundling
|
|
&& (i > required_slot
|
|
|| (required_slot == 2 && !manual_bundling_off)
|
|
|| (user_template >= 0
|
|
/* Changing from MMI to M;MI is OK. */
|
|
&& (template_val ^ required_template) > 1)))
|
|
{
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("`%s' must be last in instruction group"),
|
|
idesc->name);
|
|
if (i < 2 && required_slot == 2 && !manual_bundling_off)
|
|
manual_bundling = -1; /* Suppress meaningless post-loop errors. */
|
|
}
|
|
if (required_slot < i)
|
|
/* Can't fit this instruction. */
|
|
break;
|
|
|
|
i = required_slot;
|
|
if (required_template != template_val)
|
|
{
|
|
/* If we switch the template, we need to reset the NOPs
|
|
after slot i. The slot-types of the instructions ahead
|
|
of i never change, so we don't need to worry about
|
|
changing NOPs in front of this slot. */
|
|
for (j = i; j < 3; ++j)
|
|
insn[j] = nop[ia64_templ_desc[required_template].exec_unit[j]];
|
|
|
|
/* We just picked a template that includes the stop bit in the
|
|
middle, so we don't need another one emitted later. */
|
|
md.slot[curr].end_of_insn_group = 0;
|
|
}
|
|
template_val = required_template;
|
|
}
|
|
if (curr != first && md.slot[curr].label_fixups)
|
|
{
|
|
if (manual_bundling)
|
|
{
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("Label must be first in a bundle"));
|
|
manual_bundling = -1; /* Suppress meaningless post-loop errors. */
|
|
}
|
|
/* This insn must go into the first slot of a bundle. */
|
|
break;
|
|
}
|
|
|
|
if (end_of_insn_group && md.num_slots_in_use >= 1)
|
|
{
|
|
/* We need an instruction group boundary in the middle of a
|
|
bundle. See if we can switch to an other template with
|
|
an appropriate boundary. */
|
|
|
|
orig_template = template_val;
|
|
if (i == 1 && (user_template == 4
|
|
|| (user_template < 0
|
|
&& (ia64_templ_desc[template_val].exec_unit[0]
|
|
== IA64_UNIT_M))))
|
|
{
|
|
template_val = 5;
|
|
end_of_insn_group = 0;
|
|
}
|
|
else if (i == 2 && (user_template == 0
|
|
|| (user_template < 0
|
|
&& (ia64_templ_desc[template_val].exec_unit[1]
|
|
== IA64_UNIT_I)))
|
|
/* This test makes sure we don't switch the template if
|
|
the next instruction is one that needs to be first in
|
|
an instruction group. Since all those instructions are
|
|
in the M group, there is no way such an instruction can
|
|
fit in this bundle even if we switch the template. The
|
|
reason we have to check for this is that otherwise we
|
|
may end up generating "MI;;I M.." which has the deadly
|
|
effect that the second M instruction is no longer the
|
|
first in the group! --davidm 99/12/16 */
|
|
&& (idesc->flags & IA64_OPCODE_FIRST) == 0)
|
|
{
|
|
template_val = 1;
|
|
end_of_insn_group = 0;
|
|
}
|
|
else if (i == 1
|
|
&& user_template == 0
|
|
&& !(idesc->flags & IA64_OPCODE_FIRST))
|
|
/* Use the next slot. */
|
|
continue;
|
|
else if (curr != first)
|
|
/* can't fit this insn */
|
|
break;
|
|
|
|
if (template_val != orig_template)
|
|
/* if we switch the template, we need to reset the NOPs
|
|
after slot i. The slot-types of the instructions ahead
|
|
of i never change, so we don't need to worry about
|
|
changing NOPs in front of this slot. */
|
|
for (j = i; j < 3; ++j)
|
|
insn[j] = nop[ia64_templ_desc[template_val].exec_unit[j]];
|
|
}
|
|
required_unit = ia64_templ_desc[template_val].exec_unit[i];
|
|
|
|
/* resolve dynamic opcodes such as "break", "hint", and "nop": */
|
|
if (idesc->type == IA64_TYPE_DYN)
|
|
{
|
|
enum ia64_opnd opnd1, opnd2;
|
|
|
|
if ((strcmp (idesc->name, "nop") == 0)
|
|
|| (strcmp (idesc->name, "break") == 0))
|
|
insn_unit = required_unit;
|
|
else if (strcmp (idesc->name, "hint") == 0)
|
|
{
|
|
insn_unit = required_unit;
|
|
if (required_unit == IA64_UNIT_B)
|
|
{
|
|
switch (md.hint_b)
|
|
{
|
|
case hint_b_ok:
|
|
break;
|
|
case hint_b_warning:
|
|
as_warn (_("hint in B unit may be treated as nop"));
|
|
break;
|
|
case hint_b_error:
|
|
/* When manual bundling is off and there is no
|
|
user template, we choose a different unit so
|
|
that hint won't go into the current slot. We
|
|
will fill the current bundle with nops and
|
|
try to put hint into the next bundle. */
|
|
if (!manual_bundling && user_template < 0)
|
|
insn_unit = IA64_UNIT_I;
|
|
else
|
|
as_bad (_("hint in B unit can't be used"));
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
else if (strcmp (idesc->name, "chk.s") == 0
|
|
|| strcmp (idesc->name, "mov") == 0)
|
|
{
|
|
insn_unit = IA64_UNIT_M;
|
|
if (required_unit == IA64_UNIT_I
|
|
|| (required_unit == IA64_UNIT_F && template_val == 6))
|
|
insn_unit = IA64_UNIT_I;
|
|
}
|
|
else
|
|
as_fatal (_("emit_one_bundle: unexpected dynamic op"));
|
|
|
|
snprintf (mnemonic, sizeof (mnemonic), "%s.%c",
|
|
idesc->name, "?imbfxx"[insn_unit]);
|
|
opnd1 = idesc->operands[0];
|
|
opnd2 = idesc->operands[1];
|
|
ia64_free_opcode (idesc);
|
|
idesc = ia64_find_opcode (mnemonic);
|
|
/* moves to/from ARs have collisions */
|
|
if (opnd1 == IA64_OPND_AR3 || opnd2 == IA64_OPND_AR3)
|
|
{
|
|
while (idesc != NULL
|
|
&& (idesc->operands[0] != opnd1
|
|
|| idesc->operands[1] != opnd2))
|
|
idesc = get_next_opcode (idesc);
|
|
}
|
|
md.slot[curr].idesc = idesc;
|
|
}
|
|
else
|
|
{
|
|
insn_type = idesc->type;
|
|
insn_unit = IA64_UNIT_NIL;
|
|
switch (insn_type)
|
|
{
|
|
case IA64_TYPE_A:
|
|
if (required_unit == IA64_UNIT_I || required_unit == IA64_UNIT_M)
|
|
insn_unit = required_unit;
|
|
break;
|
|
case IA64_TYPE_X: insn_unit = IA64_UNIT_L; break;
|
|
case IA64_TYPE_I: insn_unit = IA64_UNIT_I; break;
|
|
case IA64_TYPE_M: insn_unit = IA64_UNIT_M; break;
|
|
case IA64_TYPE_B: insn_unit = IA64_UNIT_B; break;
|
|
case IA64_TYPE_F: insn_unit = IA64_UNIT_F; break;
|
|
default: break;
|
|
}
|
|
}
|
|
|
|
if (insn_unit != required_unit)
|
|
continue; /* Try next slot. */
|
|
|
|
/* Now is a good time to fix up the labels for this insn. */
|
|
mark_label = FALSE;
|
|
for (lfix = md.slot[curr].label_fixups; lfix; lfix = lfix->next)
|
|
{
|
|
S_SET_VALUE (lfix->sym, frag_now_fix () - 16);
|
|
symbol_set_frag (lfix->sym, frag_now);
|
|
mark_label |= lfix->dw2_mark_labels;
|
|
}
|
|
for (lfix = md.slot[curr].tag_fixups; lfix; lfix = lfix->next)
|
|
{
|
|
S_SET_VALUE (lfix->sym, frag_now_fix () - 16 + i);
|
|
symbol_set_frag (lfix->sym, frag_now);
|
|
}
|
|
|
|
if (debug_type == DEBUG_DWARF2
|
|
|| md.slot[curr].loc_directive_seen
|
|
|| mark_label)
|
|
{
|
|
bfd_vma addr = frag_now->fr_address + frag_now_fix () - 16 + i;
|
|
|
|
md.slot[curr].loc_directive_seen = 0;
|
|
if (mark_label)
|
|
md.slot[curr].debug_line.flags |= DWARF2_FLAG_BASIC_BLOCK;
|
|
|
|
dwarf2_gen_line_info (addr, &md.slot[curr].debug_line);
|
|
}
|
|
|
|
build_insn (md.slot + curr, insn + i);
|
|
|
|
ptr = md.slot[curr].unwind_record;
|
|
if (ptr)
|
|
{
|
|
/* Set slot numbers for all remaining unwind records belonging to the
|
|
current insn. There can not be any prologue/body unwind records
|
|
here. */
|
|
for (; ptr != end_ptr; ptr = ptr->next)
|
|
{
|
|
ptr->slot_number = (unsigned long) f + i;
|
|
ptr->slot_frag = frag_now;
|
|
}
|
|
md.slot[curr].unwind_record = NULL;
|
|
}
|
|
|
|
for (j = 0; j < md.slot[curr].num_fixups; ++j)
|
|
{
|
|
ifix = md.slot[curr].fixup + j;
|
|
fix = fix_new_exp (frag_now, frag_now_fix () - 16 + i, 8,
|
|
&ifix->expr, ifix->is_pcrel, ifix->code);
|
|
fix->tc_fix_data.opnd = ifix->opnd;
|
|
fix->fx_file = md.slot[curr].src_file;
|
|
fix->fx_line = md.slot[curr].src_line;
|
|
}
|
|
|
|
end_of_insn_group = md.slot[curr].end_of_insn_group;
|
|
|
|
/* This adjustment to "i" must occur after the fix, otherwise the fix
|
|
is assigned to the wrong slot, and the VMS linker complains. */
|
|
if (required_unit == IA64_UNIT_L)
|
|
{
|
|
know (i == 1);
|
|
/* skip one slot for long/X-unit instructions */
|
|
++i;
|
|
}
|
|
--md.num_slots_in_use;
|
|
last_slot = i;
|
|
|
|
/* clear slot: */
|
|
ia64_free_opcode (md.slot[curr].idesc);
|
|
memset (md.slot + curr, 0, sizeof (md.slot[curr]));
|
|
md.slot[curr].user_template = -1;
|
|
|
|
if (manual_bundling_off)
|
|
{
|
|
manual_bundling = 0;
|
|
break;
|
|
}
|
|
curr = (curr + 1) % NUM_SLOTS;
|
|
idesc = md.slot[curr].idesc;
|
|
}
|
|
|
|
/* A user template was specified, but the first following instruction did
|
|
not fit. This can happen with or without manual bundling. */
|
|
if (md.num_slots_in_use > 0 && last_slot < 0)
|
|
{
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("`%s' does not fit into %s template"),
|
|
idesc->name, ia64_templ_desc[template_val].name);
|
|
/* Drop first insn so we don't livelock. */
|
|
--md.num_slots_in_use;
|
|
know (curr == first);
|
|
ia64_free_opcode (md.slot[curr].idesc);
|
|
memset (md.slot + curr, 0, sizeof (md.slot[curr]));
|
|
md.slot[curr].user_template = -1;
|
|
}
|
|
else if (manual_bundling > 0)
|
|
{
|
|
if (md.num_slots_in_use > 0)
|
|
{
|
|
if (last_slot >= 2)
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("`%s' does not fit into bundle"), idesc->name);
|
|
else
|
|
{
|
|
const char *where;
|
|
|
|
if (template_val == 2)
|
|
where = "X slot";
|
|
else if (last_slot == 0)
|
|
where = "slots 2 or 3";
|
|
else
|
|
where = "slot 3";
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("`%s' can't go in %s of %s template"),
|
|
idesc->name, where, ia64_templ_desc[template_val].name);
|
|
}
|
|
}
|
|
else
|
|
as_bad_where (md.slot[curr].src_file, md.slot[curr].src_line,
|
|
_("Missing '}' at end of file"));
|
|
}
|
|
|
|
know (md.num_slots_in_use < NUM_SLOTS);
|
|
|
|
t0 = end_of_insn_group | (template_val << 1) | (insn[0] << 5) | (insn[1] << 46);
|
|
t1 = ((insn[1] >> 18) & 0x7fffff) | (insn[2] << 23);
|
|
|
|
number_to_chars_littleendian (f + 0, t0, 8);
|
|
number_to_chars_littleendian (f + 8, t1, 8);
|
|
}
|
|
|
|
int
|
|
md_parse_option (int c, const char *arg)
|
|
{
|
|
|
|
switch (c)
|
|
{
|
|
/* Switches from the Intel assembler. */
|
|
case 'm':
|
|
if (strcmp (arg, "ilp64") == 0
|
|
|| strcmp (arg, "lp64") == 0
|
|
|| strcmp (arg, "p64") == 0)
|
|
{
|
|
md.flags |= EF_IA_64_ABI64;
|
|
}
|
|
else if (strcmp (arg, "ilp32") == 0)
|
|
{
|
|
md.flags &= ~EF_IA_64_ABI64;
|
|
}
|
|
else if (strcmp (arg, "le") == 0)
|
|
{
|
|
md.flags &= ~EF_IA_64_BE;
|
|
default_big_endian = 0;
|
|
}
|
|
else if (strcmp (arg, "be") == 0)
|
|
{
|
|
md.flags |= EF_IA_64_BE;
|
|
default_big_endian = 1;
|
|
}
|
|
else if (strncmp (arg, "unwind-check=", 13) == 0)
|
|
{
|
|
arg += 13;
|
|
if (strcmp (arg, "warning") == 0)
|
|
md.unwind_check = unwind_check_warning;
|
|
else if (strcmp (arg, "error") == 0)
|
|
md.unwind_check = unwind_check_error;
|
|
else
|
|
return 0;
|
|
}
|
|
else if (strncmp (arg, "hint.b=", 7) == 0)
|
|
{
|
|
arg += 7;
|
|
if (strcmp (arg, "ok") == 0)
|
|
md.hint_b = hint_b_ok;
|
|
else if (strcmp (arg, "warning") == 0)
|
|
md.hint_b = hint_b_warning;
|
|
else if (strcmp (arg, "error") == 0)
|
|
md.hint_b = hint_b_error;
|
|
else
|
|
return 0;
|
|
}
|
|
else if (strncmp (arg, "tune=", 5) == 0)
|
|
{
|
|
arg += 5;
|
|
if (strcmp (arg, "itanium1") == 0)
|
|
md.tune = itanium1;
|
|
else if (strcmp (arg, "itanium2") == 0)
|
|
md.tune = itanium2;
|
|
else
|
|
return 0;
|
|
}
|
|
else
|
|
return 0;
|
|
break;
|
|
|
|
case 'N':
|
|
if (strcmp (arg, "so") == 0)
|
|
{
|
|
/* Suppress signon message. */
|
|
}
|
|
else if (strcmp (arg, "pi") == 0)
|
|
{
|
|
/* Reject privileged instructions. FIXME */
|
|
}
|
|
else if (strcmp (arg, "us") == 0)
|
|
{
|
|
/* Allow union of signed and unsigned range. FIXME */
|
|
}
|
|
else if (strcmp (arg, "close_fcalls") == 0)
|
|
{
|
|
/* Do not resolve global function calls. */
|
|
}
|
|
else
|
|
return 0;
|
|
break;
|
|
|
|
case 'C':
|
|
/* temp[="prefix"] Insert temporary labels into the object file
|
|
symbol table prefixed by "prefix".
|
|
Default prefix is ":temp:".
|
|
*/
|
|
break;
|
|
|
|
case 'a':
|
|
/* indirect=<tgt> Assume unannotated indirect branches behavior
|
|
according to <tgt> --
|
|
exit: branch out from the current context (default)
|
|
labels: all labels in context may be branch targets
|
|
*/
|
|
if (strncmp (arg, "indirect=", 9) != 0)
|
|
return 0;
|
|
break;
|
|
|
|
case 'x':
|
|
/* -X conflicts with an ignored option, use -x instead */
|
|
md.detect_dv = 1;
|
|
if (!arg || strcmp (arg, "explicit") == 0)
|
|
{
|
|
/* set default mode to explicit */
|
|
md.default_explicit_mode = 1;
|
|
break;
|
|
}
|
|
else if (strcmp (arg, "auto") == 0)
|
|
{
|
|
md.default_explicit_mode = 0;
|
|
}
|
|
else if (strcmp (arg, "none") == 0)
|
|
{
|
|
md.detect_dv = 0;
|
|
}
|
|
else if (strcmp (arg, "debug") == 0)
|
|
{
|
|
md.debug_dv = 1;
|
|
}
|
|
else if (strcmp (arg, "debugx") == 0)
|
|
{
|
|
md.default_explicit_mode = 1;
|
|
md.debug_dv = 1;
|
|
}
|
|
else if (strcmp (arg, "debugn") == 0)
|
|
{
|
|
md.debug_dv = 1;
|
|
md.detect_dv = 0;
|
|
}
|
|
else
|
|
{
|
|
as_bad (_("Unrecognized option '-x%s'"), arg);
|
|
}
|
|
break;
|
|
|
|
case 'S':
|
|
/* nops Print nops statistics. */
|
|
break;
|
|
|
|
/* GNU specific switches for gcc. */
|
|
case OPTION_MCONSTANT_GP:
|
|
md.flags |= EF_IA_64_CONS_GP;
|
|
break;
|
|
|
|
case OPTION_MAUTO_PIC:
|
|
md.flags |= EF_IA_64_NOFUNCDESC_CONS_GP;
|
|
break;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
void
|
|
md_show_usage (FILE *stream)
|
|
{
|
|
fputs (_("\
|
|
IA-64 options:\n\
|
|
--mconstant-gp mark output file as using the constant-GP model\n\
|
|
(sets ELF header flag EF_IA_64_CONS_GP)\n\
|
|
--mauto-pic mark output file as using the constant-GP model\n\
|
|
without function descriptors (sets ELF header flag\n\
|
|
EF_IA_64_NOFUNCDESC_CONS_GP)\n\
|
|
-milp32|-milp64|-mlp64|-mp64 select data model (default -mlp64)\n\
|
|
-mle | -mbe select little- or big-endian byte order (default -mle)\n\
|
|
-mtune=[itanium1|itanium2]\n\
|
|
tune for a specific CPU (default -mtune=itanium2)\n\
|
|
-munwind-check=[warning|error]\n\
|
|
unwind directive check (default -munwind-check=warning)\n\
|
|
-mhint.b=[ok|warning|error]\n\
|
|
hint.b check (default -mhint.b=error)\n\
|
|
-x | -xexplicit turn on dependency violation checking\n"), stream);
|
|
/* Note for translators: "automagically" can be translated as "automatically" here. */
|
|
fputs (_("\
|
|
-xauto automagically remove dependency violations (default)\n\
|
|
-xnone turn off dependency violation checking\n\
|
|
-xdebug debug dependency violation checker\n\
|
|
-xdebugn debug dependency violation checker but turn off\n\
|
|
dependency violation checking\n\
|
|
-xdebugx debug dependency violation checker and turn on\n\
|
|
dependency violation checking\n"),
|
|
stream);
|
|
}
|
|
|
|
void
|
|
ia64_after_parse_args (void)
|
|
{
|
|
if (debug_type == DEBUG_STABS)
|
|
as_fatal (_("--gstabs is not supported for ia64"));
|
|
}
|
|
|
|
/* Return true if TYPE fits in TEMPL at SLOT. */
|
|
|
|
static int
|
|
match (int templ, int type, int slot)
|
|
{
|
|
enum ia64_unit unit;
|
|
int result;
|
|
|
|
unit = ia64_templ_desc[templ].exec_unit[slot];
|
|
switch (type)
|
|
{
|
|
case IA64_TYPE_DYN: result = 1; break; /* for nop and break */
|
|
case IA64_TYPE_A:
|
|
result = (unit == IA64_UNIT_I || unit == IA64_UNIT_M);
|
|
break;
|
|
case IA64_TYPE_X: result = (unit == IA64_UNIT_L); break;
|
|
case IA64_TYPE_I: result = (unit == IA64_UNIT_I); break;
|
|
case IA64_TYPE_M: result = (unit == IA64_UNIT_M); break;
|
|
case IA64_TYPE_B: result = (unit == IA64_UNIT_B); break;
|
|
case IA64_TYPE_F: result = (unit == IA64_UNIT_F); break;
|
|
default: result = 0; break;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
/* For Itanium 1, add a bit of extra goodness if a nop of type F or B would fit
|
|
in TEMPL at SLOT. For Itanium 2, add a bit of extra goodness if a nop of
|
|
type M or I would fit in TEMPL at SLOT. */
|
|
|
|
static inline int
|
|
extra_goodness (int templ, int slot)
|
|
{
|
|
switch (md.tune)
|
|
{
|
|
case itanium1:
|
|
if (slot == 1 && match (templ, IA64_TYPE_F, slot))
|
|
return 2;
|
|
else if (slot == 2 && match (templ, IA64_TYPE_B, slot))
|
|
return 1;
|
|
else
|
|
return 0;
|
|
break;
|
|
case itanium2:
|
|
if (match (templ, IA64_TYPE_M, slot)
|
|
|| match (templ, IA64_TYPE_I, slot))
|
|
/* Favor M- and I-unit NOPs. We definitely want to avoid
|
|
F-unit and B-unit may cause split-issue or less-than-optimal
|
|
branch-prediction. */
|
|
return 2;
|
|
else
|
|
return 0;
|
|
break;
|
|
default:
|
|
abort ();
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* This function is called once, at assembler startup time. It sets
|
|
up all the tables, etc. that the MD part of the assembler will need
|
|
that can be determined before arguments are parsed. */
|
|
void
|
|
md_begin (void)
|
|
{
|
|
int i, j, k, t, goodness, best, ok;
|
|
const char *err;
|
|
char name[8];
|
|
|
|
md.auto_align = 1;
|
|
md.explicit_mode = md.default_explicit_mode;
|
|
|
|
bfd_set_section_alignment (stdoutput, text_section, 4);
|
|
|
|
/* Make sure function pointers get initialized. */
|
|
target_big_endian = -1;
|
|
dot_byteorder (default_big_endian);
|
|
|
|
alias_hash = hash_new ();
|
|
alias_name_hash = hash_new ();
|
|
secalias_hash = hash_new ();
|
|
secalias_name_hash = hash_new ();
|
|
|
|
pseudo_func[FUNC_DTP_MODULE].u.sym =
|
|
symbol_new (".<dtpmod>", undefined_section, FUNC_DTP_MODULE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_DTP_RELATIVE].u.sym =
|
|
symbol_new (".<dtprel>", undefined_section, FUNC_DTP_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_FPTR_RELATIVE].u.sym =
|
|
symbol_new (".<fptr>", undefined_section, FUNC_FPTR_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_GP_RELATIVE].u.sym =
|
|
symbol_new (".<gprel>", undefined_section, FUNC_GP_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_RELATIVE].u.sym =
|
|
symbol_new (".<ltoff>", undefined_section, FUNC_LT_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_RELATIVE_X].u.sym =
|
|
symbol_new (".<ltoffx>", undefined_section, FUNC_LT_RELATIVE_X,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_PC_RELATIVE].u.sym =
|
|
symbol_new (".<pcrel>", undefined_section, FUNC_PC_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_PLT_RELATIVE].u.sym =
|
|
symbol_new (".<pltoff>", undefined_section, FUNC_PLT_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_SEC_RELATIVE].u.sym =
|
|
symbol_new (".<secrel>", undefined_section, FUNC_SEC_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_SEG_RELATIVE].u.sym =
|
|
symbol_new (".<segrel>", undefined_section, FUNC_SEG_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_TP_RELATIVE].u.sym =
|
|
symbol_new (".<tprel>", undefined_section, FUNC_TP_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LTV_RELATIVE].u.sym =
|
|
symbol_new (".<ltv>", undefined_section, FUNC_LTV_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_FPTR_RELATIVE].u.sym =
|
|
symbol_new (".<ltoff.fptr>", undefined_section, FUNC_LT_FPTR_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_DTP_MODULE].u.sym =
|
|
symbol_new (".<ltoff.dtpmod>", undefined_section, FUNC_LT_DTP_MODULE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_DTP_RELATIVE].u.sym =
|
|
symbol_new (".<ltoff.dptrel>", undefined_section, FUNC_LT_DTP_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_LT_TP_RELATIVE].u.sym =
|
|
symbol_new (".<ltoff.tprel>", undefined_section, FUNC_LT_TP_RELATIVE,
|
|
&zero_address_frag);
|
|
|
|
pseudo_func[FUNC_IPLT_RELOC].u.sym =
|
|
symbol_new (".<iplt>", undefined_section, FUNC_IPLT_RELOC,
|
|
&zero_address_frag);
|
|
|
|
#ifdef TE_VMS
|
|
pseudo_func[FUNC_SLOTCOUNT_RELOC].u.sym =
|
|
symbol_new (".<slotcount>", undefined_section, FUNC_SLOTCOUNT_RELOC,
|
|
&zero_address_frag);
|
|
#endif
|
|
|
|
if (md.tune != itanium1)
|
|
{
|
|
/* Convert MFI NOPs bundles into MMI NOPs bundles. */
|
|
le_nop[0] = 0x8;
|
|
le_nop_stop[0] = 0x9;
|
|
}
|
|
|
|
/* Compute the table of best templates. We compute goodness as a
|
|
base 4 value, in which each match counts for 3. Match-failures
|
|
result in NOPs and we use extra_goodness() to pick the execution
|
|
units that are best suited for issuing the NOP. */
|
|
for (i = 0; i < IA64_NUM_TYPES; ++i)
|
|
for (j = 0; j < IA64_NUM_TYPES; ++j)
|
|
for (k = 0; k < IA64_NUM_TYPES; ++k)
|
|
{
|
|
best = 0;
|
|
for (t = 0; t < NELEMS (ia64_templ_desc); ++t)
|
|
{
|
|
goodness = 0;
|
|
if (match (t, i, 0))
|
|
{
|
|
if (match (t, j, 1))
|
|
{
|
|
if ((t == 2 && j == IA64_TYPE_X) || match (t, k, 2))
|
|
goodness = 3 + 3 + 3;
|
|
else
|
|
goodness = 3 + 3 + extra_goodness (t, 2);
|
|
}
|
|
else if (match (t, j, 2))
|
|
goodness = 3 + 3 + extra_goodness (t, 1);
|
|
else
|
|
{
|
|
goodness = 3;
|
|
goodness += extra_goodness (t, 1);
|
|
goodness += extra_goodness (t, 2);
|
|
}
|
|
}
|
|
else if (match (t, i, 1))
|
|
{
|
|
if ((t == 2 && i == IA64_TYPE_X) || match (t, j, 2))
|
|
goodness = 3 + 3;
|
|
else
|
|
goodness = 3 + extra_goodness (t, 2);
|
|
}
|
|
else if (match (t, i, 2))
|
|
goodness = 3 + extra_goodness (t, 1);
|
|
|
|
if (goodness > best)
|
|
{
|
|
best = goodness;
|
|
best_template[i][j][k] = t;
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifdef DEBUG_TEMPLATES
|
|
/* For debugging changes to the best_template calculations. We don't care
|
|
about combinations with invalid instructions, so start the loops at 1. */
|
|
for (i = 0; i < IA64_NUM_TYPES; ++i)
|
|
for (j = 0; j < IA64_NUM_TYPES; ++j)
|
|
for (k = 0; k < IA64_NUM_TYPES; ++k)
|
|
{
|
|
char type_letter[IA64_NUM_TYPES] = { 'n', 'a', 'i', 'm', 'b', 'f',
|
|
'x', 'd' };
|
|
fprintf (stderr, "%c%c%c %s\n", type_letter[i], type_letter[j],
|
|
type_letter[k],
|
|
ia64_templ_desc[best_template[i][j][k]].name);
|
|
}
|
|
#endif
|
|
|
|
for (i = 0; i < NUM_SLOTS; ++i)
|
|
md.slot[i].user_template = -1;
|
|
|
|
md.pseudo_hash = hash_new ();
|
|
for (i = 0; i < NELEMS (pseudo_opcode); ++i)
|
|
{
|
|
err = hash_insert (md.pseudo_hash, pseudo_opcode[i].name,
|
|
(void *) (pseudo_opcode + i));
|
|
if (err)
|
|
as_fatal (_("ia64.md_begin: can't hash `%s': %s"),
|
|
pseudo_opcode[i].name, err);
|
|
}
|
|
|
|
md.reg_hash = hash_new ();
|
|
md.dynreg_hash = hash_new ();
|
|
md.const_hash = hash_new ();
|
|
md.entry_hash = hash_new ();
|
|
|
|
/* general registers: */
|
|
declare_register_set ("r", 128, REG_GR);
|
|
declare_register ("gp", REG_GR + 1);
|
|
declare_register ("sp", REG_GR + 12);
|
|
declare_register ("tp", REG_GR + 13);
|
|
declare_register_set ("ret", 4, REG_GR + 8);
|
|
|
|
/* floating point registers: */
|
|
declare_register_set ("f", 128, REG_FR);
|
|
declare_register_set ("farg", 8, REG_FR + 8);
|
|
declare_register_set ("fret", 8, REG_FR + 8);
|
|
|
|
/* branch registers: */
|
|
declare_register_set ("b", 8, REG_BR);
|
|
declare_register ("rp", REG_BR + 0);
|
|
|
|
/* predicate registers: */
|
|
declare_register_set ("p", 64, REG_P);
|
|
declare_register ("pr", REG_PR);
|
|
declare_register ("pr.rot", REG_PR_ROT);
|
|
|
|
/* application registers: */
|
|
declare_register_set ("ar", 128, REG_AR);
|
|
for (i = 0; i < NELEMS (ar); ++i)
|
|
declare_register (ar[i].name, REG_AR + ar[i].regnum);
|
|
|
|
/* control registers: */
|
|
declare_register_set ("cr", 128, REG_CR);
|
|
for (i = 0; i < NELEMS (cr); ++i)
|
|
declare_register (cr[i].name, REG_CR + cr[i].regnum);
|
|
|
|
/* dahr registers: */
|
|
declare_register_set ("dahr", 8, REG_DAHR);
|
|
|
|
declare_register ("ip", REG_IP);
|
|
declare_register ("cfm", REG_CFM);
|
|
declare_register ("psr", REG_PSR);
|
|
declare_register ("psr.l", REG_PSR_L);
|
|
declare_register ("psr.um", REG_PSR_UM);
|
|
|
|
for (i = 0; i < NELEMS (indirect_reg); ++i)
|
|
{
|
|
unsigned int regnum = indirect_reg[i].regnum;
|
|
|
|
md.indregsym[regnum - IND_CPUID] = declare_register (indirect_reg[i].name, regnum);
|
|
}
|
|
|
|
/* pseudo-registers used to specify unwind info: */
|
|
declare_register ("psp", REG_PSP);
|
|
|
|
for (i = 0; i < NELEMS (const_bits); ++i)
|
|
{
|
|
err = hash_insert (md.const_hash, const_bits[i].name,
|
|
(void *) (const_bits + i));
|
|
if (err)
|
|
as_fatal (_("Inserting \"%s\" into constant hash table failed: %s"),
|
|
name, err);
|
|
}
|
|
|
|
/* Set the architecture and machine depending on defaults and command line
|
|
options. */
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
ok = bfd_set_arch_mach (stdoutput, bfd_arch_ia64, bfd_mach_ia64_elf64);
|
|
else
|
|
ok = bfd_set_arch_mach (stdoutput, bfd_arch_ia64, bfd_mach_ia64_elf32);
|
|
|
|
if (! ok)
|
|
as_warn (_("Could not set architecture and machine"));
|
|
|
|
/* Set the pointer size and pointer shift size depending on md.flags */
|
|
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
{
|
|
md.pointer_size = 8; /* pointers are 8 bytes */
|
|
md.pointer_size_shift = 3; /* alignment is 8 bytes = 2^2 */
|
|
}
|
|
else
|
|
{
|
|
md.pointer_size = 4; /* pointers are 4 bytes */
|
|
md.pointer_size_shift = 2; /* alignment is 4 bytes = 2^2 */
|
|
}
|
|
|
|
md.mem_offset.hint = 0;
|
|
md.path = 0;
|
|
md.maxpaths = 0;
|
|
md.entry_labels = NULL;
|
|
}
|
|
|
|
/* Set the default options in md. Cannot do this in md_begin because
|
|
that is called after md_parse_option which is where we set the
|
|
options in md based on command line options. */
|
|
|
|
void
|
|
ia64_init (int argc ATTRIBUTE_UNUSED, char **argv ATTRIBUTE_UNUSED)
|
|
{
|
|
md.flags = MD_FLAGS_DEFAULT;
|
|
#ifndef TE_VMS
|
|
/* Don't turn on dependency checking for VMS, doesn't work. */
|
|
md.detect_dv = 1;
|
|
#endif
|
|
/* FIXME: We should change it to unwind_check_error someday. */
|
|
md.unwind_check = unwind_check_warning;
|
|
md.hint_b = hint_b_error;
|
|
md.tune = itanium2;
|
|
}
|
|
|
|
/* Return a string for the target object file format. */
|
|
|
|
const char *
|
|
ia64_target_format (void)
|
|
{
|
|
if (OUTPUT_FLAVOR == bfd_target_elf_flavour)
|
|
{
|
|
if (md.flags & EF_IA_64_BE)
|
|
{
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
#if defined(TE_AIX50)
|
|
return "elf64-ia64-aix-big";
|
|
#elif defined(TE_HPUX)
|
|
return "elf64-ia64-hpux-big";
|
|
#else
|
|
return "elf64-ia64-big";
|
|
#endif
|
|
else
|
|
#if defined(TE_AIX50)
|
|
return "elf32-ia64-aix-big";
|
|
#elif defined(TE_HPUX)
|
|
return "elf32-ia64-hpux-big";
|
|
#else
|
|
return "elf32-ia64-big";
|
|
#endif
|
|
}
|
|
else
|
|
{
|
|
if (md.flags & EF_IA_64_ABI64)
|
|
#if defined (TE_AIX50)
|
|
return "elf64-ia64-aix-little";
|
|
#elif defined (TE_VMS)
|
|
{
|
|
md.flags |= EF_IA_64_ARCHVER_1;
|
|
return "elf64-ia64-vms";
|
|
}
|
|
#else
|
|
return "elf64-ia64-little";
|
|
#endif
|
|
else
|
|
#ifdef TE_AIX50
|
|
return "elf32-ia64-aix-little";
|
|
#else
|
|
return "elf32-ia64-little";
|
|
#endif
|
|
}
|
|
}
|
|
else
|
|
return "unknown-format";
|
|
}
|
|
|
|
void
|
|
ia64_end_of_source (void)
|
|
{
|
|
/* terminate insn group upon reaching end of file: */
|
|
insn_group_break (1, 0, 0);
|
|
|
|
/* emits slots we haven't written yet: */
|
|
ia64_flush_insns ();
|
|
|
|
bfd_set_private_flags (stdoutput, md.flags);
|
|
|
|
md.mem_offset.hint = 0;
|
|
}
|
|
|
|
void
|
|
ia64_start_line (void)
|
|
{
|
|
static int first;
|
|
|
|
if (!first) {
|
|
/* Make sure we don't reference input_line_pointer[-1] when that's
|
|
not valid. */
|
|
first = 1;
|
|
return;
|
|
}
|
|
|
|
if (md.qp.X_op == O_register)
|
|
as_bad (_("qualifying predicate not followed by instruction"));
|
|
md.qp.X_op = O_absent;
|
|
|
|
if (ignore_input ())
|
|
return;
|
|
|
|
if (input_line_pointer[0] == ';' && input_line_pointer[-1] == ';')
|
|
{
|
|
if (md.detect_dv && !md.explicit_mode)
|
|
{
|
|
static int warned;
|
|
|
|
if (!warned)
|
|
{
|
|
warned = 1;
|
|
as_warn (_("Explicit stops are ignored in auto mode"));
|
|
}
|
|
}
|
|
else
|
|
insn_group_break (1, 0, 0);
|
|
}
|
|
else if (input_line_pointer[-1] == '{')
|
|
{
|
|
if (md.manual_bundling)
|
|
as_warn (_("Found '{' when manual bundling is already turned on"));
|
|
else
|
|
CURR_SLOT.manual_bundling_on = 1;
|
|
md.manual_bundling = 1;
|
|
|
|
/* Bundling is only acceptable in explicit mode
|
|
or when in default automatic mode. */
|
|
if (md.detect_dv && !md.explicit_mode)
|
|
{
|
|
if (!md.mode_explicitly_set
|
|
&& !md.default_explicit_mode)
|
|
dot_dv_mode ('E');
|
|
else
|
|
as_warn (_("Found '{' after explicit switch to automatic mode"));
|
|
}
|
|
}
|
|
else if (input_line_pointer[-1] == '}')
|
|
{
|
|
if (!md.manual_bundling)
|
|
as_warn (_("Found '}' when manual bundling is off"));
|
|
else
|
|
PREV_SLOT.manual_bundling_off = 1;
|
|
md.manual_bundling = 0;
|
|
|
|
/* switch back to automatic mode, if applicable */
|
|
if (md.detect_dv
|
|
&& md.explicit_mode
|
|
&& !md.mode_explicitly_set
|
|
&& !md.default_explicit_mode)
|
|
dot_dv_mode ('A');
|
|
}
|
|
}
|
|
|
|
/* This is a hook for ia64_frob_label, so that it can distinguish tags from
|
|
labels. */
|
|
static int defining_tag = 0;
|
|
|
|
int
|
|
ia64_unrecognized_line (int ch)
|
|
{
|
|
switch (ch)
|
|
{
|
|
case '(':
|
|
expression_and_evaluate (&md.qp);
|
|
if (*input_line_pointer++ != ')')
|
|
{
|
|
as_bad (_("Expected ')'"));
|
|
return 0;
|
|
}
|
|
if (md.qp.X_op != O_register)
|
|
{
|
|
as_bad (_("Qualifying predicate expected"));
|
|
return 0;
|
|
}
|
|
if (md.qp.X_add_number < REG_P || md.qp.X_add_number >= REG_P + 64)
|
|
{
|
|
as_bad (_("Predicate register expected"));
|
|
return 0;
|
|
}
|
|
return 1;
|
|
|
|
case '[':
|
|
{
|
|
char *s;
|
|
char c;
|
|
symbolS *tag;
|
|
int temp;
|
|
|
|
if (md.qp.X_op == O_register)
|
|
{
|
|
as_bad (_("Tag must come before qualifying predicate."));
|
|
return 0;
|
|
}
|
|
|
|
/* This implements just enough of read_a_source_file in read.c to
|
|
recognize labels. */
|
|
if (is_name_beginner (*input_line_pointer))
|
|
{
|
|
c = get_symbol_name (&s);
|
|
}
|
|
else if (LOCAL_LABELS_FB
|
|
&& ISDIGIT (*input_line_pointer))
|
|
{
|
|
temp = 0;
|
|
while (ISDIGIT (*input_line_pointer))
|
|
temp = (temp * 10) + *input_line_pointer++ - '0';
|
|
fb_label_instance_inc (temp);
|
|
s = fb_label_name (temp, 0);
|
|
c = *input_line_pointer;
|
|
}
|
|
else
|
|
{
|
|
s = NULL;
|
|
c = '\0';
|
|
}
|
|
if (c != ':')
|
|
{
|
|
/* Put ':' back for error messages' sake. */
|
|
*input_line_pointer++ = ':';
|
|
as_bad (_("Expected ':'"));
|
|
return 0;
|
|
}
|
|
|
|
defining_tag = 1;
|
|
tag = colon (s);
|
|
defining_tag = 0;
|
|
/* Put ':' back for error messages' sake. */
|
|
*input_line_pointer++ = ':';
|
|
if (*input_line_pointer++ != ']')
|
|
{
|
|
as_bad (_("Expected ']'"));
|
|
return 0;
|
|
}
|
|
if (! tag)
|
|
{
|
|
as_bad (_("Tag name expected"));
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* Not a valid line. */
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
ia64_frob_label (struct symbol *sym)
|
|
{
|
|
struct label_fix *fix;
|
|
|
|
/* Tags need special handling since they are not bundle breaks like
|
|
labels. */
|
|
if (defining_tag)
|
|
{
|
|
fix = XOBNEW (¬es, struct label_fix);
|
|
fix->sym = sym;
|
|
fix->next = CURR_SLOT.tag_fixups;
|
|
fix->dw2_mark_labels = FALSE;
|
|
CURR_SLOT.tag_fixups = fix;
|
|
|
|
return;
|
|
}
|
|
|
|
if (bfd_get_section_flags (stdoutput, now_seg) & SEC_CODE)
|
|
{
|
|
md.last_text_seg = now_seg;
|
|
fix = XOBNEW (¬es, struct label_fix);
|
|
fix->sym = sym;
|
|
fix->next = CURR_SLOT.label_fixups;
|
|
fix->dw2_mark_labels = dwarf2_loc_mark_labels;
|
|
CURR_SLOT.label_fixups = fix;
|
|
|
|
/* Keep track of how many code entry points we've seen. */
|
|
if (md.path == md.maxpaths)
|
|
{
|
|
md.maxpaths += 20;
|
|
md.entry_labels = XRESIZEVEC (const char *, md.entry_labels,
|
|
md.maxpaths);
|
|
}
|
|
md.entry_labels[md.path++] = S_GET_NAME (sym);
|
|
}
|
|
}
|
|
|
|
#ifdef TE_HPUX
|
|
/* The HP-UX linker will give unresolved symbol errors for symbols
|
|
that are declared but unused. This routine removes declared,
|
|
unused symbols from an object. */
|
|
int
|
|
ia64_frob_symbol (struct symbol *sym)
|
|
{
|
|
if ((S_GET_SEGMENT (sym) == bfd_und_section_ptr && ! symbol_used_p (sym) &&
|
|
ELF_ST_VISIBILITY (S_GET_OTHER (sym)) == STV_DEFAULT)
|
|
|| (S_GET_SEGMENT (sym) == bfd_abs_section_ptr
|
|
&& ! S_IS_EXTERNAL (sym)))
|
|
return 1;
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
void
|
|
ia64_flush_pending_output (void)
|
|
{
|
|
if (!md.keep_pending_output
|
|
&& bfd_get_section_flags (stdoutput, now_seg) & SEC_CODE)
|
|
{
|
|
/* ??? This causes many unnecessary stop bits to be emitted.
|
|
Unfortunately, it isn't clear if it is safe to remove this. */
|
|
insn_group_break (1, 0, 0);
|
|
ia64_flush_insns ();
|
|
}
|
|
}
|
|
|
|
/* Do ia64-specific expression optimization. All that's done here is
|
|
to transform index expressions that are either due to the indexing
|
|
of rotating registers or due to the indexing of indirect register
|
|
sets. */
|
|
int
|
|
ia64_optimize_expr (expressionS *l, operatorT op, expressionS *r)
|
|
{
|
|
if (op != O_index)
|
|
return 0;
|
|
resolve_expression (l);
|
|
if (l->X_op == O_register)
|
|
{
|
|
unsigned num_regs = l->X_add_number >> 16;
|
|
|
|
resolve_expression (r);
|
|
if (num_regs)
|
|
{
|
|
/* Left side is a .rotX-allocated register. */
|
|
if (r->X_op != O_constant)
|
|
{
|
|
as_bad (_("Rotating register index must be a non-negative constant"));
|
|
r->X_add_number = 0;
|
|
}
|
|
else if ((valueT) r->X_add_number >= num_regs)
|
|
{
|
|
as_bad (_("Index out of range 0..%u"), num_regs - 1);
|
|
r->X_add_number = 0;
|
|
}
|
|
l->X_add_number = (l->X_add_number & 0xffff) + r->X_add_number;
|
|
return 1;
|
|
}
|
|
else if (l->X_add_number >= IND_CPUID && l->X_add_number <= IND_RR)
|
|
{
|
|
if (r->X_op != O_register
|
|
|| r->X_add_number < REG_GR
|
|
|| r->X_add_number > REG_GR + 127)
|
|
{
|
|
as_bad (_("Indirect register index must be a general register"));
|
|
r->X_add_number = REG_GR;
|
|
}
|
|
l->X_op = O_index;
|
|
l->X_op_symbol = md.indregsym[l->X_add_number - IND_CPUID];
|
|
l->X_add_number = r->X_add_number;
|
|
return 1;
|
|
}
|
|
}
|
|
as_bad (_("Index can only be applied to rotating or indirect registers"));
|
|
/* Fall back to some register use of which has as little as possible
|
|
side effects, to minimize subsequent error messages. */
|
|
l->X_op = O_register;
|
|
l->X_add_number = REG_GR + 3;
|
|
return 1;
|
|
}
|
|
|
|
int
|
|
ia64_parse_name (char *name, expressionS *e, char *nextcharP)
|
|
{
|
|
struct const_desc *cdesc;
|
|
struct dynreg *dr = 0;
|
|
unsigned int idx;
|
|
struct symbol *sym;
|
|
char *end;
|
|
|
|
if (*name == '@')
|
|
{
|
|
enum pseudo_type pseudo_type = PSEUDO_FUNC_NONE;
|
|
|
|
/* Find what relocation pseudo-function we're dealing with. */
|
|
for (idx = 0; idx < NELEMS (pseudo_func); ++idx)
|
|
if (pseudo_func[idx].name
|
|
&& pseudo_func[idx].name[0] == name[1]
|
|
&& strcmp (pseudo_func[idx].name + 1, name + 2) == 0)
|
|
{
|
|
pseudo_type = pseudo_func[idx].type;
|
|
break;
|
|
}
|
|
switch (pseudo_type)
|
|
{
|
|
case PSEUDO_FUNC_RELOC:
|
|
end = input_line_pointer;
|
|
if (*nextcharP != '(')
|
|
{
|
|
as_bad (_("Expected '('"));
|
|
break;
|
|
}
|
|
/* Skip '('. */
|
|
++input_line_pointer;
|
|
expression (e);
|
|
if (*input_line_pointer != ')')
|
|
{
|
|
as_bad (_("Missing ')'"));
|
|
goto done;
|
|
}
|
|
/* Skip ')'. */
|
|
++input_line_pointer;
|
|
#ifdef TE_VMS
|
|
if (idx == FUNC_SLOTCOUNT_RELOC)
|
|
{
|
|
/* @slotcount can accept any expression. Canonicalize. */
|
|
e->X_add_symbol = make_expr_symbol (e);
|
|
e->X_op = O_symbol;
|
|
e->X_add_number = 0;
|
|
}
|
|
#endif
|
|
if (e->X_op != O_symbol)
|
|
{
|
|
if (e->X_op != O_pseudo_fixup)
|
|
{
|
|
as_bad (_("Not a symbolic expression"));
|
|
goto done;
|
|
}
|
|
if (idx != FUNC_LT_RELATIVE)
|
|
{
|
|
as_bad (_("Illegal combination of relocation functions"));
|
|
goto done;
|
|
}
|
|
switch (S_GET_VALUE (e->X_op_symbol))
|
|
{
|
|
case FUNC_FPTR_RELATIVE:
|
|
idx = FUNC_LT_FPTR_RELATIVE; break;
|
|
case FUNC_DTP_MODULE:
|
|
idx = FUNC_LT_DTP_MODULE; break;
|
|
case FUNC_DTP_RELATIVE:
|
|
idx = FUNC_LT_DTP_RELATIVE; break;
|
|
case FUNC_TP_RELATIVE:
|
|
idx = FUNC_LT_TP_RELATIVE; break;
|
|
default:
|
|
as_bad (_("Illegal combination of relocation functions"));
|
|
goto done;
|
|
}
|
|
}
|
|
/* Make sure gas doesn't get rid of local symbols that are used
|
|
in relocs. */
|
|
e->X_op = O_pseudo_fixup;
|
|
e->X_op_symbol = pseudo_func[idx].u.sym;
|
|
done:
|
|
*nextcharP = *input_line_pointer;
|
|
break;
|
|
|
|
case PSEUDO_FUNC_CONST:
|
|
e->X_op = O_constant;
|
|
e->X_add_number = pseudo_func[idx].u.ival;
|
|
break;
|
|
|
|
case PSEUDO_FUNC_REG:
|
|
e->X_op = O_register;
|
|
e->X_add_number = pseudo_func[idx].u.ival;
|
|
break;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/* first see if NAME is a known register name: */
|
|
sym = hash_find (md.reg_hash, name);
|
|
if (sym)
|
|
{
|
|
e->X_op = O_register;
|
|
e->X_add_number = S_GET_VALUE (sym);
|
|
return 1;
|
|
}
|
|
|
|
cdesc = hash_find (md.const_hash, name);
|
|
if (cdesc)
|
|
{
|
|
e->X_op = O_constant;
|
|
e->X_add_number = cdesc->value;
|
|
return 1;
|
|
}
|
|
|
|
/* check for inN, locN, or outN: */
|
|
idx = 0;
|
|
switch (name[0])
|
|
{
|
|
case 'i':
|
|
if (name[1] == 'n' && ISDIGIT (name[2]))
|
|
{
|
|
dr = &md.in;
|
|
idx = 2;
|
|
}
|
|
break;
|
|
|
|
case 'l':
|
|
if (name[1] == 'o' && name[2] == 'c' && ISDIGIT (name[3]))
|
|
{
|
|
dr = &md.loc;
|
|
idx = 3;
|
|
}
|
|
break;
|
|
|
|
case 'o':
|
|
if (name[1] == 'u' && name[2] == 't' && ISDIGIT (name[3]))
|
|
{
|
|
dr = &md.out;
|
|
idx = 3;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* Ignore register numbers with leading zeroes, except zero itself. */
|
|
if (dr && (name[idx] != '0' || name[idx + 1] == '\0'))
|
|
{
|
|
unsigned long regnum;
|
|
|
|
/* The name is inN, locN, or outN; parse the register number. */
|
|
regnum = strtoul (name + idx, &end, 10);
|
|
if (end > name + idx && *end == '\0' && regnum < 96)
|
|
{
|
|
if (regnum >= dr->num_regs)
|
|
{
|
|
if (!dr->num_regs)
|
|
as_bad (_("No current frame"));
|
|
else
|
|
as_bad (_("Register number out of range 0..%u"),
|
|
dr->num_regs - 1);
|
|
regnum = 0;
|
|
}
|
|
e->X_op = O_register;
|
|
e->X_add_number = dr->base + regnum;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
end = xstrdup (name);
|
|
name = ia64_canonicalize_symbol_name (end);
|
|
if ((dr = hash_find (md.dynreg_hash, name)))
|
|
{
|
|
/* We've got ourselves the name of a rotating register set.
|
|
Store the base register number in the low 16 bits of
|
|
X_add_number and the size of the register set in the top 16
|
|
bits. */
|
|
e->X_op = O_register;
|
|
e->X_add_number = dr->base | (dr->num_regs << 16);
|
|
free (end);
|
|
return 1;
|
|
}
|
|
free (end);
|
|
return 0;
|
|
}
|
|
|
|
/* Remove the '#' suffix that indicates a symbol as opposed to a register. */
|
|
|
|
char *
|
|
ia64_canonicalize_symbol_name (char *name)
|
|
{
|
|
size_t len = strlen (name), full = len;
|
|
|
|
while (len > 0 && name[len - 1] == '#')
|
|
--len;
|
|
if (len <= 0)
|
|
{
|
|
if (full > 0)
|
|
as_bad (_("Standalone `#' is illegal"));
|
|
}
|
|
else if (len < full - 1)
|
|
as_warn (_("Redundant `#' suffix operators"));
|
|
name[len] = '\0';
|
|
return name;
|
|
}
|
|
|
|
/* Return true if idesc is a conditional branch instruction. This excludes
|
|
the modulo scheduled branches, and br.ia. Mod-sched branches are excluded
|
|
because they always read/write resources regardless of the value of the
|
|
qualifying predicate. br.ia must always use p0, and hence is always
|
|
taken. Thus this function returns true for branches which can fall
|
|
through, and which use no resources if they do fall through. */
|
|
|
|
static int
|
|
is_conditional_branch (struct ia64_opcode *idesc)
|
|
{
|
|
/* br is a conditional branch. Everything that starts with br. except
|
|
br.ia, br.c{loop,top,exit}, and br.w{top,exit} is a conditional branch.
|
|
Everything that starts with brl is a conditional branch. */
|
|
return (idesc->name[0] == 'b' && idesc->name[1] == 'r'
|
|
&& (idesc->name[2] == '\0'
|
|
|| (idesc->name[2] == '.' && idesc->name[3] != 'i'
|
|
&& idesc->name[3] != 'c' && idesc->name[3] != 'w')
|
|
|| idesc->name[2] == 'l'
|
|
/* br.cond, br.call, br.clr */
|
|
|| (idesc->name[2] == '.' && idesc->name[3] == 'c'
|
|
&& (idesc->name[4] == 'a' || idesc->name[4] == 'o'
|
|
|| (idesc->name[4] == 'l' && idesc->name[5] == 'r')))));
|
|
}
|
|
|
|
/* Return whether the given opcode is a taken branch. If there's any doubt,
|
|
returns zero. */
|
|
|
|
static int
|
|
is_taken_branch (struct ia64_opcode *idesc)
|
|
{
|
|
return ((is_conditional_branch (idesc) && CURR_SLOT.qp_regno == 0)
|
|
|| strncmp (idesc->name, "br.ia", 5) == 0);
|
|
}
|
|
|
|
/* Return whether the given opcode is an interruption or rfi. If there's any
|
|
doubt, returns zero. */
|
|
|
|
static int
|
|
is_interruption_or_rfi (struct ia64_opcode *idesc)
|
|
{
|
|
if (strcmp (idesc->name, "rfi") == 0)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
/* Returns the index of the given dependency in the opcode's list of chks, or
|
|
-1 if there is no dependency. */
|
|
|
|
static int
|
|
depends_on (int depind, struct ia64_opcode *idesc)
|
|
{
|
|
int i;
|
|
const struct ia64_opcode_dependency *dep = idesc->dependencies;
|
|
for (i = 0; i < dep->nchks; i++)
|
|
{
|
|
if (depind == DEP (dep->chks[i]))
|
|
return i;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/* Determine a set of specific resources used for a particular resource
|
|
class. Returns the number of specific resources identified For those
|
|
cases which are not determinable statically, the resource returned is
|
|
marked nonspecific.
|
|
|
|
Meanings of value in 'NOTE':
|
|
1) only read/write when the register number is explicitly encoded in the
|
|
insn.
|
|
2) only read CFM when accessing a rotating GR, FR, or PR. mov pr only
|
|
accesses CFM when qualifying predicate is in the rotating region.
|
|
3) general register value is used to specify an indirect register; not
|
|
determinable statically.
|
|
4) only read the given resource when bits 7:0 of the indirect index
|
|
register value does not match the register number of the resource; not
|
|
determinable statically.
|
|
5) all rules are implementation specific.
|
|
6) only when both the index specified by the reader and the index specified
|
|
by the writer have the same value in bits 63:61; not determinable
|
|
statically.
|
|
7) only access the specified resource when the corresponding mask bit is
|
|
set
|
|
8) PSR.dfh is only read when these insns reference FR32-127. PSR.dfl is
|
|
only read when these insns reference FR2-31
|
|
9) PSR.mfl is only written when these insns write FR2-31. PSR.mfh is only
|
|
written when these insns write FR32-127
|
|
10) The PSR.bn bit is only accessed when one of GR16-31 is specified in the
|
|
instruction
|
|
11) The target predicates are written independently of PR[qp], but source
|
|
registers are only read if PR[qp] is true. Since the state of PR[qp]
|
|
cannot statically be determined, all source registers are marked used.
|
|
12) This insn only reads the specified predicate register when that
|
|
register is the PR[qp].
|
|
13) This reference to ld-c only applies to the GR whose value is loaded
|
|
with data returned from memory, not the post-incremented address register.
|
|
14) The RSE resource includes the implementation-specific RSE internal
|
|
state resources. At least one (and possibly more) of these resources are
|
|
read by each instruction listed in IC:rse-readers. At least one (and
|
|
possibly more) of these resources are written by each insn listed in
|
|
IC:rse-writers.
|
|
15+16) Represents reserved instructions, which the assembler does not
|
|
generate.
|
|
17) CR[TPR] has a RAW dependency only between mov-to-CR-TPR and
|
|
mov-to-PSR-l or ssm instructions that set PSR.i, PSR.pp or PSR.up.
|
|
|
|
Memory resources (i.e. locations in memory) are *not* marked or tracked by
|
|
this code; there are no dependency violations based on memory access.
|
|
*/
|
|
|
|
#define MAX_SPECS 256
|
|
#define DV_CHK 1
|
|
#define DV_REG 0
|
|
|
|
static int
|
|
specify_resource (const struct ia64_dependency *dep,
|
|
struct ia64_opcode *idesc,
|
|
/* is this a DV chk or a DV reg? */
|
|
int type,
|
|
/* returned specific resources */
|
|
struct rsrc specs[MAX_SPECS],
|
|
/* resource note for this insn's usage */
|
|
int note,
|
|
/* which execution path to examine */
|
|
int path)
|
|
{
|
|
int count = 0;
|
|
int i;
|
|
int rsrc_write = 0;
|
|
struct rsrc tmpl;
|
|
|
|
if (dep->mode == IA64_DV_WAW
|
|
|| (dep->mode == IA64_DV_RAW && type == DV_REG)
|
|
|| (dep->mode == IA64_DV_WAR && type == DV_CHK))
|
|
rsrc_write = 1;
|
|
|
|
/* template for any resources we identify */
|
|
tmpl.dependency = dep;
|
|
tmpl.note = note;
|
|
tmpl.insn_srlz = tmpl.data_srlz = 0;
|
|
tmpl.qp_regno = CURR_SLOT.qp_regno;
|
|
tmpl.link_to_qp_branch = 1;
|
|
tmpl.mem_offset.hint = 0;
|
|
tmpl.mem_offset.offset = 0;
|
|
tmpl.mem_offset.base = 0;
|
|
tmpl.specific = 1;
|
|
tmpl.index = -1;
|
|
tmpl.cmp_type = CMP_NONE;
|
|
tmpl.depind = 0;
|
|
tmpl.file = NULL;
|
|
tmpl.line = 0;
|
|
tmpl.path = 0;
|
|
|
|
#define UNHANDLED \
|
|
as_warn (_("Unhandled dependency %s for %s (%s), note %d"), \
|
|
dep->name, idesc->name, (rsrc_write?"write":"read"), note)
|
|
#define KNOWN(REG) (gr_values[REG].known && gr_values[REG].path >= path)
|
|
|
|
/* we don't need to track these */
|
|
if (dep->semantics == IA64_DVS_NONE)
|
|
return 0;
|
|
|
|
switch (dep->specifier)
|
|
{
|
|
case IA64_RS_AR_K:
|
|
if (note == 1)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if (regno >= 0 && regno <= 7)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 0)
|
|
{
|
|
for (i = 0; i < 8; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_AR_UNAT:
|
|
/* This is a mov =AR or mov AR= instruction. */
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if (regno == AR_UNAT)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* This is a spill/fill, or other instruction that modifies the
|
|
unat register. */
|
|
|
|
/* Unless we can determine the specific bits used, mark the whole
|
|
thing; bits 8:3 of the memory address indicate the bit used in
|
|
UNAT. The .mem.offset hint may be used to eliminate a small
|
|
subset of conflicts. */
|
|
specs[count] = tmpl;
|
|
if (md.mem_offset.hint)
|
|
{
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Using hint for spill/fill\n");
|
|
/* The index isn't actually used, just set it to something
|
|
approximating the bit index. */
|
|
specs[count].index = (md.mem_offset.offset >> 3) & 0x3F;
|
|
specs[count].mem_offset.hint = 1;
|
|
specs[count].mem_offset.offset = md.mem_offset.offset;
|
|
specs[count++].mem_offset.base = md.mem_offset.base;
|
|
}
|
|
else
|
|
{
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_AR:
|
|
if (note == 1)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if ((regno >= 8 && regno <= 15)
|
|
|| (regno >= 20 && regno <= 23)
|
|
|| (regno >= 31 && regno <= 39)
|
|
|| (regno >= 41 && regno <= 47)
|
|
|| (regno >= 67 && regno <= 111))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_ARb:
|
|
if (note == 1)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if ((regno >= 48 && regno <= 63)
|
|
|| (regno >= 112 && regno <= 127))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 0)
|
|
{
|
|
for (i = 48; i < 64; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
for (i = 112; i < 128; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_BR:
|
|
if (note != 1)
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
else
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
if (idesc->operands[i] == IA64_OPND_B1
|
|
|| idesc->operands[i] == IA64_OPND_B2)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_BR;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
for (i = idesc->num_outputs; i < NELEMS (idesc->operands); i++)
|
|
if (idesc->operands[i] == IA64_OPND_B1
|
|
|| idesc->operands[i] == IA64_OPND_B2)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_BR;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CPUID: /* four or more registers */
|
|
if (note == 3)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CPUID_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_DBR: /* four or more registers */
|
|
if (note == 3)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_DBR_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 0 && !rsrc_write)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_IBR: /* four or more registers */
|
|
if (note == 3)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_IBR_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_MSR:
|
|
if (note == 5)
|
|
{
|
|
/* These are implementation specific. Force all references to
|
|
conflict with all other references. */
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_PKR: /* 16 or more registers */
|
|
if (note == 3 || note == 4)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_PKR_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
if (note == 3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
for (i = 0; i < NELEMS (gr_values); i++)
|
|
{
|
|
/* Uses all registers *except* the one in R3. */
|
|
if ((unsigned)i != (gr_values[regno].value & 0xFF))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 0)
|
|
{
|
|
/* probe et al. */
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_PMC: /* four or more registers */
|
|
if (note == 3)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_PMC_R3
|
|
|| (!rsrc_write && idesc->operands[1] == IA64_OPND_PMD_R3))
|
|
|
|
{
|
|
int reg_index = ((idesc->operands[1] == IA64_OPND_R3 && !rsrc_write)
|
|
? 1 : !rsrc_write);
|
|
int regno = CURR_SLOT.opnd[reg_index].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_PMD: /* four or more registers */
|
|
if (note == 3)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_PMD_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = gr_values[regno].value & 0xFF;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_RR: /* eight registers */
|
|
if (note == 6)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_RR_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_GR;
|
|
if (regno >= 0 && regno < NELEMS (gr_values)
|
|
&& KNOWN (regno))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = (gr_values[regno].value >> 61) & 0x7;
|
|
}
|
|
else
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 0 && !rsrc_write)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CR_IRR:
|
|
if (note == 0)
|
|
{
|
|
/* handle mov-from-CR-IVR; it's a read that writes CR[IRR] */
|
|
int regno = CURR_SLOT.opnd[1].X_add_number - REG_CR;
|
|
if (rsrc_write
|
|
&& idesc->operands[1] == IA64_OPND_CR3
|
|
&& regno == CR_IVR)
|
|
{
|
|
for (i = 0; i < 4; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CR_IRR0 + i;
|
|
}
|
|
}
|
|
}
|
|
else if (note == 1)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_CR;
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CR3
|
|
&& regno >= CR_IRR0
|
|
&& regno <= CR_IRR3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CR_IIB:
|
|
if (note != 0)
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
else
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_CR;
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CR3
|
|
&& (regno == CR_IIB0 || regno == CR_IIB1))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CR_LRR:
|
|
if (note != 1)
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
else
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_CR;
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CR3
|
|
&& (regno == CR_LRR0 || regno == CR_LRR1))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CR:
|
|
if (note == 1)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CR3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_CR;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_DAHR:
|
|
if (note == 0)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_DAHR3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_DAHR;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_FR:
|
|
case IA64_RS_FRb:
|
|
if (note != 1)
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
else if (rsrc_write)
|
|
{
|
|
if (dep->specifier == IA64_RS_FRb
|
|
&& idesc->operands[0] == IA64_OPND_F1)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.opnd[0].X_add_number - REG_FR;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
for (i = idesc->num_outputs; i < NELEMS (idesc->operands); i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_F2
|
|
|| idesc->operands[i] == IA64_OPND_F3
|
|
|| idesc->operands[i] == IA64_OPND_F4)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_FR;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_GR:
|
|
if (note == 13)
|
|
{
|
|
/* This reference applies only to the GR whose value is loaded with
|
|
data returned from memory. */
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.opnd[0].X_add_number - REG_GR;
|
|
}
|
|
else if (note == 1)
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
if (idesc->operands[i] == IA64_OPND_R1
|
|
|| idesc->operands[i] == IA64_OPND_R2
|
|
|| idesc->operands[i] == IA64_OPND_R3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
}
|
|
if (idesc->flags & IA64_OPCODE_POSTINC)
|
|
for (i = 0; i < NELEMS (idesc->operands); i++)
|
|
if (idesc->operands[i] == IA64_OPND_MR3)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* Look for anything that reads a GR. */
|
|
for (i = 0; i < NELEMS (idesc->operands); i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_MR3
|
|
|| idesc->operands[i] == IA64_OPND_CPUID_R3
|
|
|| idesc->operands[i] == IA64_OPND_DBR_R3
|
|
|| idesc->operands[i] == IA64_OPND_IBR_R3
|
|
|| idesc->operands[i] == IA64_OPND_MSR_R3
|
|
|| idesc->operands[i] == IA64_OPND_PKR_R3
|
|
|| idesc->operands[i] == IA64_OPND_PMC_R3
|
|
|| idesc->operands[i] == IA64_OPND_PMD_R3
|
|
|| idesc->operands[i] == IA64_OPND_DAHR_R3
|
|
|| idesc->operands[i] == IA64_OPND_RR_R3
|
|
|| ((i >= idesc->num_outputs)
|
|
&& (idesc->operands[i] == IA64_OPND_R1
|
|
|| idesc->operands[i] == IA64_OPND_R2
|
|
|| idesc->operands[i] == IA64_OPND_R3
|
|
/* addl source register. */
|
|
|| idesc->operands[i] == IA64_OPND_R3_2)))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index =
|
|
CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
/* This is the same as IA64_RS_PRr, except that the register range is
|
|
from 1 - 15, and there are no rotating register reads/writes here. */
|
|
case IA64_RS_PR:
|
|
if (note == 0)
|
|
{
|
|
for (i = 1; i < 16; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else if (note == 7)
|
|
{
|
|
valueT mask = 0;
|
|
/* Mark only those registers indicated by the mask. */
|
|
if (rsrc_write)
|
|
{
|
|
mask = CURR_SLOT.opnd[2].X_add_number;
|
|
for (i = 1; i < 16; i++)
|
|
if (mask & ((valueT) 1 << i))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
}
|
|
else if (note == 11) /* note 11 implies note 1 as well */
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_P1
|
|
|| idesc->operands[i] == IA64_OPND_P2)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_P;
|
|
if (regno >= 1 && regno < 16)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
}
|
|
else if (note == 12)
|
|
{
|
|
if (CURR_SLOT.qp_regno >= 1 && CURR_SLOT.qp_regno < 16)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.qp_regno;
|
|
}
|
|
}
|
|
else if (note == 1)
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
int p1 = CURR_SLOT.opnd[0].X_add_number - REG_P;
|
|
int p2 = CURR_SLOT.opnd[1].X_add_number - REG_P;
|
|
int or_andcm = strstr (idesc->name, "or.andcm") != NULL;
|
|
int and_orcm = strstr (idesc->name, "and.orcm") != NULL;
|
|
|
|
if ((idesc->operands[0] == IA64_OPND_P1
|
|
|| idesc->operands[0] == IA64_OPND_P2)
|
|
&& p1 >= 1 && p1 < 16)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count].cmp_type =
|
|
(or_andcm ? CMP_OR : (and_orcm ? CMP_AND : CMP_NONE));
|
|
specs[count++].index = p1;
|
|
}
|
|
if ((idesc->operands[1] == IA64_OPND_P1
|
|
|| idesc->operands[1] == IA64_OPND_P2)
|
|
&& p2 >= 1 && p2 < 16)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count].cmp_type =
|
|
(or_andcm ? CMP_AND : (and_orcm ? CMP_OR : CMP_NONE));
|
|
specs[count++].index = p2;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (CURR_SLOT.qp_regno >= 1 && CURR_SLOT.qp_regno < 16)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.qp_regno;
|
|
}
|
|
if (idesc->operands[1] == IA64_OPND_PR)
|
|
{
|
|
for (i = 1; i < 16; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
/* This is the general case for PRs. IA64_RS_PR and IA64_RS_PR63 are
|
|
simplified cases of this. */
|
|
case IA64_RS_PRr:
|
|
if (note == 0)
|
|
{
|
|
for (i = 16; i < 63; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else if (note == 7)
|
|
{
|
|
valueT mask = 0;
|
|
/* Mark only those registers indicated by the mask. */
|
|
if (rsrc_write
|
|
&& idesc->operands[0] == IA64_OPND_PR)
|
|
{
|
|
mask = CURR_SLOT.opnd[2].X_add_number;
|
|
if (mask & ((valueT) 1 << 16))
|
|
for (i = 16; i < 63; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else if (rsrc_write
|
|
&& idesc->operands[0] == IA64_OPND_PR_ROT)
|
|
{
|
|
for (i = 16; i < 63; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
}
|
|
else if (note == 11) /* note 11 implies note 1 as well */
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_P1
|
|
|| idesc->operands[i] == IA64_OPND_P2)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_P;
|
|
if (regno >= 16 && regno < 63)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = regno;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
}
|
|
else if (note == 12)
|
|
{
|
|
if (CURR_SLOT.qp_regno >= 16 && CURR_SLOT.qp_regno < 63)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.qp_regno;
|
|
}
|
|
}
|
|
else if (note == 1)
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
int p1 = CURR_SLOT.opnd[0].X_add_number - REG_P;
|
|
int p2 = CURR_SLOT.opnd[1].X_add_number - REG_P;
|
|
int or_andcm = strstr (idesc->name, "or.andcm") != NULL;
|
|
int and_orcm = strstr (idesc->name, "and.orcm") != NULL;
|
|
|
|
if ((idesc->operands[0] == IA64_OPND_P1
|
|
|| idesc->operands[0] == IA64_OPND_P2)
|
|
&& p1 >= 16 && p1 < 63)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count].cmp_type =
|
|
(or_andcm ? CMP_OR : (and_orcm ? CMP_AND : CMP_NONE));
|
|
specs[count++].index = p1;
|
|
}
|
|
if ((idesc->operands[1] == IA64_OPND_P1
|
|
|| idesc->operands[1] == IA64_OPND_P2)
|
|
&& p2 >= 16 && p2 < 63)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count].cmp_type =
|
|
(or_andcm ? CMP_AND : (and_orcm ? CMP_OR : CMP_NONE));
|
|
specs[count++].index = p2;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (CURR_SLOT.qp_regno >= 16 && CURR_SLOT.qp_regno < 63)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = CURR_SLOT.qp_regno;
|
|
}
|
|
if (idesc->operands[1] == IA64_OPND_PR)
|
|
{
|
|
for (i = 16; i < 63; i++)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].index = i;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_PSR:
|
|
/* Verify that the instruction is using the PSR bit indicated in
|
|
dep->regindex. */
|
|
if (note == 0)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_PSR_UM)
|
|
{
|
|
if (dep->regindex < 6)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (idesc->operands[!rsrc_write] == IA64_OPND_PSR)
|
|
{
|
|
if (dep->regindex < 32
|
|
|| dep->regindex == 35
|
|
|| dep->regindex == 36
|
|
|| (!rsrc_write && dep->regindex == PSR_CPL))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (idesc->operands[!rsrc_write] == IA64_OPND_PSR_L)
|
|
{
|
|
if (dep->regindex < 32
|
|
|| dep->regindex == 35
|
|
|| dep->regindex == 36
|
|
|| (rsrc_write && dep->regindex == PSR_CPL))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* Several PSR bits have very specific dependencies. */
|
|
switch (dep->regindex)
|
|
{
|
|
default:
|
|
specs[count++] = tmpl;
|
|
break;
|
|
case PSR_IC:
|
|
if (rsrc_write)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else
|
|
{
|
|
/* Only certain CR accesses use PSR.ic */
|
|
if (idesc->operands[0] == IA64_OPND_CR3
|
|
|| idesc->operands[1] == IA64_OPND_CR3)
|
|
{
|
|
int reg_index =
|
|
((idesc->operands[0] == IA64_OPND_CR3)
|
|
? 0 : 1);
|
|
int regno =
|
|
CURR_SLOT.opnd[reg_index].X_add_number - REG_CR;
|
|
|
|
switch (regno)
|
|
{
|
|
default:
|
|
break;
|
|
case CR_ITIR:
|
|
case CR_IFS:
|
|
case CR_IIM:
|
|
case CR_IIP:
|
|
case CR_IPSR:
|
|
case CR_ISR:
|
|
case CR_IFA:
|
|
case CR_IHA:
|
|
case CR_IIB0:
|
|
case CR_IIB1:
|
|
case CR_IIPA:
|
|
specs[count++] = tmpl;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case PSR_CPL:
|
|
if (rsrc_write)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else
|
|
{
|
|
/* Only some AR accesses use cpl */
|
|
if (idesc->operands[0] == IA64_OPND_AR3
|
|
|| idesc->operands[1] == IA64_OPND_AR3)
|
|
{
|
|
int reg_index =
|
|
((idesc->operands[0] == IA64_OPND_AR3)
|
|
? 0 : 1);
|
|
int regno =
|
|
CURR_SLOT.opnd[reg_index].X_add_number - REG_AR;
|
|
|
|
if (regno == AR_ITC
|
|
|| regno == AR_RUC
|
|
|| (reg_index == 0
|
|
&& (regno == AR_RSC
|
|
|| (regno >= AR_K0
|
|
&& regno <= AR_K7))))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if (note == 7)
|
|
{
|
|
valueT mask = 0;
|
|
if (idesc->operands[0] == IA64_OPND_IMMU24)
|
|
{
|
|
mask = CURR_SLOT.opnd[0].X_add_number;
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
if (mask & ((valueT) 1 << dep->regindex))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (note == 8)
|
|
{
|
|
int min = dep->regindex == PSR_DFL ? 2 : 32;
|
|
int max = dep->regindex == PSR_DFL ? 31 : 127;
|
|
/* dfh is read on FR32-127; dfl is read on FR2-31 */
|
|
for (i = 0; i < NELEMS (idesc->operands); i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_F1
|
|
|| idesc->operands[i] == IA64_OPND_F2
|
|
|| idesc->operands[i] == IA64_OPND_F3
|
|
|| idesc->operands[i] == IA64_OPND_F4)
|
|
{
|
|
int reg = CURR_SLOT.opnd[i].X_add_number - REG_FR;
|
|
if (reg >= min && reg <= max)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if (note == 9)
|
|
{
|
|
int min = dep->regindex == PSR_MFL ? 2 : 32;
|
|
int max = dep->regindex == PSR_MFL ? 31 : 127;
|
|
/* mfh is read on writes to FR32-127; mfl is read on writes to
|
|
FR2-31 */
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_F1)
|
|
{
|
|
int reg = CURR_SLOT.opnd[i].X_add_number - REG_FR;
|
|
if (reg >= min && reg <= max)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if (note == 10)
|
|
{
|
|
for (i = 0; i < NELEMS (idesc->operands); i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_R1
|
|
|| idesc->operands[i] == IA64_OPND_R2
|
|
|| idesc->operands[i] == IA64_OPND_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
if (regno >= 16 && regno <= 31)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_AR_FPSR:
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if (regno == AR_FPSR)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_ARX:
|
|
/* Handle all AR[REG] resources */
|
|
if (note == 0 || note == 1)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_AR;
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_AR3
|
|
&& regno == dep->regindex)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
/* other AR[REG] resources may be affected by AR accesses */
|
|
else if (idesc->operands[0] == IA64_OPND_AR3)
|
|
{
|
|
/* AR[] writes */
|
|
regno = CURR_SLOT.opnd[0].X_add_number - REG_AR;
|
|
switch (dep->regindex)
|
|
{
|
|
default:
|
|
break;
|
|
case AR_BSP:
|
|
case AR_RNAT:
|
|
if (regno == AR_BSPSTORE)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
/* Fall through. */
|
|
case AR_RSC:
|
|
if (!rsrc_write &&
|
|
(regno == AR_BSPSTORE
|
|
|| regno == AR_RNAT))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
else if (idesc->operands[1] == IA64_OPND_AR3)
|
|
{
|
|
/* AR[] reads */
|
|
regno = CURR_SLOT.opnd[1].X_add_number - REG_AR;
|
|
switch (dep->regindex)
|
|
{
|
|
default:
|
|
break;
|
|
case AR_RSC:
|
|
if (regno == AR_BSPSTORE || regno == AR_RNAT)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CRX:
|
|
/* Handle all CR[REG] resources.
|
|
??? FIXME: The rule 17 isn't really handled correctly. */
|
|
if (note == 0 || note == 1 || note == 17)
|
|
{
|
|
if (idesc->operands[!rsrc_write] == IA64_OPND_CR3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[!rsrc_write].X_add_number - REG_CR;
|
|
if (regno == dep->regindex)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else if (!rsrc_write)
|
|
{
|
|
/* Reads from CR[IVR] affect other resources. */
|
|
if (regno == CR_IVR)
|
|
{
|
|
if ((dep->regindex >= CR_IRR0
|
|
&& dep->regindex <= CR_IRR3)
|
|
|| dep->regindex == CR_TPR)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_INSERVICE:
|
|
/* look for write of EOI (67) or read of IVR (65) */
|
|
if ((idesc->operands[0] == IA64_OPND_CR3
|
|
&& CURR_SLOT.opnd[0].X_add_number - REG_CR == CR_EOI)
|
|
|| (idesc->operands[1] == IA64_OPND_CR3
|
|
&& CURR_SLOT.opnd[1].X_add_number - REG_CR == CR_IVR))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_GR0:
|
|
if (note == 1)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_CFM:
|
|
if (note != 2)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else
|
|
{
|
|
/* Check if any of the registers accessed are in the rotating region.
|
|
mov to/from pr accesses CFM only when qp_regno is in the rotating
|
|
region */
|
|
for (i = 0; i < NELEMS (idesc->operands); i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_R1
|
|
|| idesc->operands[i] == IA64_OPND_R2
|
|
|| idesc->operands[i] == IA64_OPND_R3)
|
|
{
|
|
int num = CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
/* Assumes that md.rot.num_regs is always valid */
|
|
if (md.rot.num_regs > 0
|
|
&& num > 31
|
|
&& num < 31 + md.rot.num_regs)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_F1
|
|
|| idesc->operands[i] == IA64_OPND_F2
|
|
|| idesc->operands[i] == IA64_OPND_F3
|
|
|| idesc->operands[i] == IA64_OPND_F4)
|
|
{
|
|
int num = CURR_SLOT.opnd[i].X_add_number - REG_FR;
|
|
if (num > 31)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_P1
|
|
|| idesc->operands[i] == IA64_OPND_P2)
|
|
{
|
|
int num = CURR_SLOT.opnd[i].X_add_number - REG_P;
|
|
if (num > 15)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
}
|
|
if (CURR_SLOT.qp_regno > 15)
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].specific = 0;
|
|
}
|
|
}
|
|
break;
|
|
|
|
/* This is the same as IA64_RS_PRr, except simplified to account for
|
|
the fact that there is only one register. */
|
|
case IA64_RS_PR63:
|
|
if (note == 0)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
else if (note == 7)
|
|
{
|
|
valueT mask = 0;
|
|
if (idesc->operands[2] == IA64_OPND_IMM17)
|
|
mask = CURR_SLOT.opnd[2].X_add_number;
|
|
if (mask & ((valueT) 1 << 63))
|
|
specs[count++] = tmpl;
|
|
}
|
|
else if (note == 11)
|
|
{
|
|
if ((idesc->operands[0] == IA64_OPND_P1
|
|
&& CURR_SLOT.opnd[0].X_add_number - REG_P == 63)
|
|
|| (idesc->operands[1] == IA64_OPND_P2
|
|
&& CURR_SLOT.opnd[1].X_add_number - REG_P == 63))
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (note == 12)
|
|
{
|
|
if (CURR_SLOT.qp_regno == 63)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (note == 1)
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
int p1 = CURR_SLOT.opnd[0].X_add_number - REG_P;
|
|
int p2 = CURR_SLOT.opnd[1].X_add_number - REG_P;
|
|
int or_andcm = strstr (idesc->name, "or.andcm") != NULL;
|
|
int and_orcm = strstr (idesc->name, "and.orcm") != NULL;
|
|
|
|
if (p1 == 63
|
|
&& (idesc->operands[0] == IA64_OPND_P1
|
|
|| idesc->operands[0] == IA64_OPND_P2))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].cmp_type =
|
|
(or_andcm ? CMP_OR : (and_orcm ? CMP_AND : CMP_NONE));
|
|
}
|
|
if (p2 == 63
|
|
&& (idesc->operands[1] == IA64_OPND_P1
|
|
|| idesc->operands[1] == IA64_OPND_P2))
|
|
{
|
|
specs[count] = tmpl;
|
|
specs[count++].cmp_type =
|
|
(or_andcm ? CMP_AND : (and_orcm ? CMP_OR : CMP_NONE));
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (CURR_SLOT.qp_regno == 63)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
UNHANDLED;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_RSE:
|
|
/* FIXME we can identify some individual RSE written resources, but RSE
|
|
read resources have not yet been completely identified, so for now
|
|
treat RSE as a single resource */
|
|
if (strncmp (idesc->name, "mov", 3) == 0)
|
|
{
|
|
if (rsrc_write)
|
|
{
|
|
if (idesc->operands[0] == IA64_OPND_AR3
|
|
&& CURR_SLOT.opnd[0].X_add_number - REG_AR == AR_BSPSTORE)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (idesc->operands[0] == IA64_OPND_AR3)
|
|
{
|
|
if (CURR_SLOT.opnd[0].X_add_number - REG_AR == AR_BSPSTORE
|
|
|| CURR_SLOT.opnd[0].X_add_number - REG_AR == AR_RNAT)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
else if (idesc->operands[1] == IA64_OPND_AR3)
|
|
{
|
|
if (CURR_SLOT.opnd[1].X_add_number - REG_AR == AR_BSP
|
|
|| CURR_SLOT.opnd[1].X_add_number - REG_AR == AR_BSPSTORE
|
|
|| CURR_SLOT.opnd[1].X_add_number - REG_AR == AR_RNAT)
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
specs[count++] = tmpl;
|
|
}
|
|
break;
|
|
|
|
case IA64_RS_ANY:
|
|
/* FIXME -- do any of these need to be non-specific? */
|
|
specs[count++] = tmpl;
|
|
break;
|
|
|
|
default:
|
|
as_bad (_("Unrecognized dependency specifier %d\n"), dep->specifier);
|
|
break;
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
/* Clear branch flags on marked resources. This breaks the link between the
|
|
QP of the marking instruction and a subsequent branch on the same QP. */
|
|
|
|
static void
|
|
clear_qp_branch_flag (valueT mask)
|
|
{
|
|
int i;
|
|
for (i = 0; i < regdepslen; i++)
|
|
{
|
|
valueT bit = ((valueT) 1 << regdeps[i].qp_regno);
|
|
if ((bit & mask) != 0)
|
|
{
|
|
regdeps[i].link_to_qp_branch = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* MASK contains 2 and only 2 PRs which are mutually exclusive. Remove
|
|
any mutexes which contain one of the PRs and create new ones when
|
|
needed. */
|
|
|
|
static int
|
|
update_qp_mutex (valueT mask)
|
|
{
|
|
int i;
|
|
int add = 0;
|
|
|
|
i = 0;
|
|
while (i < qp_mutexeslen)
|
|
{
|
|
if ((qp_mutexes[i].prmask & mask) != 0)
|
|
{
|
|
/* If it destroys and creates the same mutex, do nothing. */
|
|
if (qp_mutexes[i].prmask == mask
|
|
&& qp_mutexes[i].path == md.path)
|
|
{
|
|
i++;
|
|
add = -1;
|
|
}
|
|
else
|
|
{
|
|
int keep = 0;
|
|
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Clearing mutex relation");
|
|
print_prmask (qp_mutexes[i].prmask);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
|
|
/* Deal with the old mutex with more than 3+ PRs only if
|
|
the new mutex on the same execution path with it.
|
|
|
|
FIXME: The 3+ mutex support is incomplete.
|
|
dot_pred_rel () may be a better place to fix it. */
|
|
if (qp_mutexes[i].path == md.path)
|
|
{
|
|
/* If it is a proper subset of the mutex, create a
|
|
new mutex. */
|
|
if (add == 0
|
|
&& (qp_mutexes[i].prmask & mask) == mask)
|
|
add = 1;
|
|
|
|
qp_mutexes[i].prmask &= ~mask;
|
|
if (qp_mutexes[i].prmask & (qp_mutexes[i].prmask - 1))
|
|
{
|
|
/* Modify the mutex if there are more than one
|
|
PR left. */
|
|
keep = 1;
|
|
i++;
|
|
}
|
|
}
|
|
|
|
if (keep == 0)
|
|
/* Remove the mutex. */
|
|
qp_mutexes[i] = qp_mutexes[--qp_mutexeslen];
|
|
}
|
|
}
|
|
else
|
|
++i;
|
|
}
|
|
|
|
if (add == 1)
|
|
add_qp_mutex (mask);
|
|
|
|
return add;
|
|
}
|
|
|
|
/* Remove any mutexes which contain any of the PRs indicated in the mask.
|
|
|
|
Any changes to a PR clears the mutex relations which include that PR. */
|
|
|
|
static void
|
|
clear_qp_mutex (valueT mask)
|
|
{
|
|
int i;
|
|
|
|
i = 0;
|
|
while (i < qp_mutexeslen)
|
|
{
|
|
if ((qp_mutexes[i].prmask & mask) != 0)
|
|
{
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Clearing mutex relation");
|
|
print_prmask (qp_mutexes[i].prmask);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
qp_mutexes[i] = qp_mutexes[--qp_mutexeslen];
|
|
}
|
|
else
|
|
++i;
|
|
}
|
|
}
|
|
|
|
/* Clear implies relations which contain PRs in the given masks.
|
|
P1_MASK indicates the source of the implies relation, while P2_MASK
|
|
indicates the implied PR. */
|
|
|
|
static void
|
|
clear_qp_implies (valueT p1_mask, valueT p2_mask)
|
|
{
|
|
int i;
|
|
|
|
i = 0;
|
|
while (i < qp_implieslen)
|
|
{
|
|
if ((((valueT) 1 << qp_implies[i].p1) & p1_mask) != 0
|
|
|| (((valueT) 1 << qp_implies[i].p2) & p2_mask) != 0)
|
|
{
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Clearing implied relation PR%d->PR%d\n",
|
|
qp_implies[i].p1, qp_implies[i].p2);
|
|
qp_implies[i] = qp_implies[--qp_implieslen];
|
|
}
|
|
else
|
|
++i;
|
|
}
|
|
}
|
|
|
|
/* Add the PRs specified to the list of implied relations. */
|
|
|
|
static void
|
|
add_qp_imply (int p1, int p2)
|
|
{
|
|
valueT mask;
|
|
valueT bit;
|
|
int i;
|
|
|
|
/* p0 is not meaningful here. */
|
|
if (p1 == 0 || p2 == 0)
|
|
abort ();
|
|
|
|
if (p1 == p2)
|
|
return;
|
|
|
|
/* If it exists already, ignore it. */
|
|
for (i = 0; i < qp_implieslen; i++)
|
|
{
|
|
if (qp_implies[i].p1 == p1
|
|
&& qp_implies[i].p2 == p2
|
|
&& qp_implies[i].path == md.path
|
|
&& !qp_implies[i].p2_branched)
|
|
return;
|
|
}
|
|
|
|
if (qp_implieslen == qp_impliestotlen)
|
|
{
|
|
qp_impliestotlen += 20;
|
|
qp_implies = XRESIZEVEC (struct qp_imply, qp_implies, qp_impliestotlen);
|
|
}
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Registering PR%d implies PR%d\n", p1, p2);
|
|
qp_implies[qp_implieslen].p1 = p1;
|
|
qp_implies[qp_implieslen].p2 = p2;
|
|
qp_implies[qp_implieslen].path = md.path;
|
|
qp_implies[qp_implieslen++].p2_branched = 0;
|
|
|
|
/* Add in the implied transitive relations; for everything that p2 implies,
|
|
make p1 imply that, too; for everything that implies p1, make it imply p2
|
|
as well. */
|
|
for (i = 0; i < qp_implieslen; i++)
|
|
{
|
|
if (qp_implies[i].p1 == p2)
|
|
add_qp_imply (p1, qp_implies[i].p2);
|
|
if (qp_implies[i].p2 == p1)
|
|
add_qp_imply (qp_implies[i].p1, p2);
|
|
}
|
|
/* Add in mutex relations implied by this implies relation; for each mutex
|
|
relation containing p2, duplicate it and replace p2 with p1. */
|
|
bit = (valueT) 1 << p1;
|
|
mask = (valueT) 1 << p2;
|
|
for (i = 0; i < qp_mutexeslen; i++)
|
|
{
|
|
if (qp_mutexes[i].prmask & mask)
|
|
add_qp_mutex ((qp_mutexes[i].prmask & ~mask) | bit);
|
|
}
|
|
}
|
|
|
|
/* Add the PRs specified in the mask to the mutex list; this means that only
|
|
one of the PRs can be true at any time. PR0 should never be included in
|
|
the mask. */
|
|
|
|
static void
|
|
add_qp_mutex (valueT mask)
|
|
{
|
|
if (mask & 0x1)
|
|
abort ();
|
|
|
|
if (qp_mutexeslen == qp_mutexestotlen)
|
|
{
|
|
qp_mutexestotlen += 20;
|
|
qp_mutexes = XRESIZEVEC (struct qpmutex, qp_mutexes, qp_mutexestotlen);
|
|
}
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Registering mutex on");
|
|
print_prmask (mask);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
qp_mutexes[qp_mutexeslen].path = md.path;
|
|
qp_mutexes[qp_mutexeslen++].prmask = mask;
|
|
}
|
|
|
|
static int
|
|
has_suffix_p (const char *name, const char *suffix)
|
|
{
|
|
size_t namelen = strlen (name);
|
|
size_t sufflen = strlen (suffix);
|
|
|
|
if (namelen <= sufflen)
|
|
return 0;
|
|
return strcmp (name + namelen - sufflen, suffix) == 0;
|
|
}
|
|
|
|
static void
|
|
clear_register_values (void)
|
|
{
|
|
int i;
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Clearing register values\n");
|
|
for (i = 1; i < NELEMS (gr_values); i++)
|
|
gr_values[i].known = 0;
|
|
}
|
|
|
|
/* Keep track of register values/changes which affect DV tracking.
|
|
|
|
optimization note: should add a flag to classes of insns where otherwise we
|
|
have to examine a group of strings to identify them. */
|
|
|
|
static void
|
|
note_register_values (struct ia64_opcode *idesc)
|
|
{
|
|
valueT qp_changemask = 0;
|
|
int i;
|
|
|
|
/* Invalidate values for registers being written to. */
|
|
for (i = 0; i < idesc->num_outputs; i++)
|
|
{
|
|
if (idesc->operands[i] == IA64_OPND_R1
|
|
|| idesc->operands[i] == IA64_OPND_R2
|
|
|| idesc->operands[i] == IA64_OPND_R3)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
if (regno > 0 && regno < NELEMS (gr_values))
|
|
gr_values[regno].known = 0;
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_R3_2)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_GR;
|
|
if (regno > 0 && regno < 4)
|
|
gr_values[regno].known = 0;
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_P1
|
|
|| idesc->operands[i] == IA64_OPND_P2)
|
|
{
|
|
int regno = CURR_SLOT.opnd[i].X_add_number - REG_P;
|
|
qp_changemask |= (valueT) 1 << regno;
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_PR)
|
|
{
|
|
if (idesc->operands[2] & (valueT) 0x10000)
|
|
qp_changemask = ~(valueT) 0x1FFFF | idesc->operands[2];
|
|
else
|
|
qp_changemask = idesc->operands[2];
|
|
break;
|
|
}
|
|
else if (idesc->operands[i] == IA64_OPND_PR_ROT)
|
|
{
|
|
if (idesc->operands[1] & ((valueT) 1 << 43))
|
|
qp_changemask = -((valueT) 1 << 44) | idesc->operands[1];
|
|
else
|
|
qp_changemask = idesc->operands[1];
|
|
qp_changemask &= ~(valueT) 0xFFFF;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Always clear qp branch flags on any PR change. */
|
|
/* FIXME there may be exceptions for certain compares. */
|
|
clear_qp_branch_flag (qp_changemask);
|
|
|
|
/* Invalidate rotating registers on insns which affect RRBs in CFM. */
|
|
if (idesc->flags & IA64_OPCODE_MOD_RRBS)
|
|
{
|
|
qp_changemask |= ~(valueT) 0xFFFF;
|
|
if (strcmp (idesc->name, "clrrrb.pr") != 0)
|
|
{
|
|
for (i = 32; i < 32 + md.rot.num_regs; i++)
|
|
gr_values[i].known = 0;
|
|
}
|
|
clear_qp_mutex (qp_changemask);
|
|
clear_qp_implies (qp_changemask, qp_changemask);
|
|
}
|
|
/* After a call, all register values are undefined, except those marked
|
|
as "safe". */
|
|
else if (strncmp (idesc->name, "br.call", 6) == 0
|
|
|| strncmp (idesc->name, "brl.call", 7) == 0)
|
|
{
|
|
/* FIXME keep GR values which are marked as "safe_across_calls" */
|
|
clear_register_values ();
|
|
clear_qp_mutex (~qp_safe_across_calls);
|
|
clear_qp_implies (~qp_safe_across_calls, ~qp_safe_across_calls);
|
|
clear_qp_branch_flag (~qp_safe_across_calls);
|
|
}
|
|
else if (is_interruption_or_rfi (idesc)
|
|
|| is_taken_branch (idesc))
|
|
{
|
|
clear_register_values ();
|
|
clear_qp_mutex (~(valueT) 0);
|
|
clear_qp_implies (~(valueT) 0, ~(valueT) 0);
|
|
}
|
|
/* Look for mutex and implies relations. */
|
|
else if ((idesc->operands[0] == IA64_OPND_P1
|
|
|| idesc->operands[0] == IA64_OPND_P2)
|
|
&& (idesc->operands[1] == IA64_OPND_P1
|
|
|| idesc->operands[1] == IA64_OPND_P2))
|
|
{
|
|
int p1 = CURR_SLOT.opnd[0].X_add_number - REG_P;
|
|
int p2 = CURR_SLOT.opnd[1].X_add_number - REG_P;
|
|
valueT p1mask = (p1 != 0) ? (valueT) 1 << p1 : 0;
|
|
valueT p2mask = (p2 != 0) ? (valueT) 1 << p2 : 0;
|
|
|
|
/* If both PRs are PR0, we can't really do anything. */
|
|
if (p1 == 0 && p2 == 0)
|
|
{
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Ignoring PRs due to inclusion of p0\n");
|
|
}
|
|
/* In general, clear mutexes and implies which include P1 or P2,
|
|
with the following exceptions. */
|
|
else if (has_suffix_p (idesc->name, ".or.andcm")
|
|
|| has_suffix_p (idesc->name, ".and.orcm"))
|
|
{
|
|
clear_qp_implies (p2mask, p1mask);
|
|
}
|
|
else if (has_suffix_p (idesc->name, ".andcm")
|
|
|| has_suffix_p (idesc->name, ".and"))
|
|
{
|
|
clear_qp_implies (0, p1mask | p2mask);
|
|
}
|
|
else if (has_suffix_p (idesc->name, ".orcm")
|
|
|| has_suffix_p (idesc->name, ".or"))
|
|
{
|
|
clear_qp_mutex (p1mask | p2mask);
|
|
clear_qp_implies (p1mask | p2mask, 0);
|
|
}
|
|
else
|
|
{
|
|
int added = 0;
|
|
|
|
clear_qp_implies (p1mask | p2mask, p1mask | p2mask);
|
|
|
|
/* If one of the PRs is PR0, we call clear_qp_mutex. */
|
|
if (p1 == 0 || p2 == 0)
|
|
clear_qp_mutex (p1mask | p2mask);
|
|
else
|
|
added = update_qp_mutex (p1mask | p2mask);
|
|
|
|
if (CURR_SLOT.qp_regno == 0
|
|
|| has_suffix_p (idesc->name, ".unc"))
|
|
{
|
|
if (added == 0 && p1 && p2)
|
|
add_qp_mutex (p1mask | p2mask);
|
|
if (CURR_SLOT.qp_regno != 0)
|
|
{
|
|
if (p1)
|
|
add_qp_imply (p1, CURR_SLOT.qp_regno);
|
|
if (p2)
|
|
add_qp_imply (p2, CURR_SLOT.qp_regno);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
/* Look for mov imm insns into GRs. */
|
|
else if (idesc->operands[0] == IA64_OPND_R1
|
|
&& (idesc->operands[1] == IA64_OPND_IMM22
|
|
|| idesc->operands[1] == IA64_OPND_IMMU64)
|
|
&& CURR_SLOT.opnd[1].X_op == O_constant
|
|
&& (strcmp (idesc->name, "mov") == 0
|
|
|| strcmp (idesc->name, "movl") == 0))
|
|
{
|
|
int regno = CURR_SLOT.opnd[0].X_add_number - REG_GR;
|
|
if (regno > 0 && regno < NELEMS (gr_values))
|
|
{
|
|
gr_values[regno].known = 1;
|
|
gr_values[regno].value = CURR_SLOT.opnd[1].X_add_number;
|
|
gr_values[regno].path = md.path;
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Know gr%d = ", regno);
|
|
fprintf_vma (stderr, gr_values[regno].value);
|
|
fputs ("\n", stderr);
|
|
}
|
|
}
|
|
}
|
|
/* Look for dep.z imm insns. */
|
|
else if (idesc->operands[0] == IA64_OPND_R1
|
|
&& idesc->operands[1] == IA64_OPND_IMM8
|
|
&& strcmp (idesc->name, "dep.z") == 0)
|
|
{
|
|
int regno = CURR_SLOT.opnd[0].X_add_number - REG_GR;
|
|
if (regno > 0 && regno < NELEMS (gr_values))
|
|
{
|
|
valueT value = CURR_SLOT.opnd[1].X_add_number;
|
|
|
|
if (CURR_SLOT.opnd[3].X_add_number < 64)
|
|
value &= ((valueT)1 << CURR_SLOT.opnd[3].X_add_number) - 1;
|
|
value <<= CURR_SLOT.opnd[2].X_add_number;
|
|
gr_values[regno].known = 1;
|
|
gr_values[regno].value = value;
|
|
gr_values[regno].path = md.path;
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Know gr%d = ", regno);
|
|
fprintf_vma (stderr, gr_values[regno].value);
|
|
fputs ("\n", stderr);
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
clear_qp_mutex (qp_changemask);
|
|
clear_qp_implies (qp_changemask, qp_changemask);
|
|
}
|
|
}
|
|
|
|
/* Return whether the given predicate registers are currently mutex. */
|
|
|
|
static int
|
|
qp_mutex (int p1, int p2, int path)
|
|
{
|
|
int i;
|
|
valueT mask;
|
|
|
|
if (p1 != p2)
|
|
{
|
|
mask = ((valueT) 1 << p1) | (valueT) 1 << p2;
|
|
for (i = 0; i < qp_mutexeslen; i++)
|
|
{
|
|
if (qp_mutexes[i].path >= path
|
|
&& (qp_mutexes[i].prmask & mask) == mask)
|
|
return 1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Return whether the given resource is in the given insn's list of chks
|
|
Return 1 if the conflict is absolutely determined, 2 if it's a potential
|
|
conflict. */
|
|
|
|
static int
|
|
resources_match (struct rsrc *rs,
|
|
struct ia64_opcode *idesc,
|
|
int note,
|
|
int qp_regno,
|
|
int path)
|
|
{
|
|
struct rsrc specs[MAX_SPECS];
|
|
int count;
|
|
|
|
/* If the marked resource's qp_regno and the given qp_regno are mutex,
|
|
we don't need to check. One exception is note 11, which indicates that
|
|
target predicates are written regardless of PR[qp]. */
|
|
if (qp_mutex (rs->qp_regno, qp_regno, path)
|
|
&& note != 11)
|
|
return 0;
|
|
|
|
count = specify_resource (rs->dependency, idesc, DV_CHK, specs, note, path);
|
|
while (count-- > 0)
|
|
{
|
|
/* UNAT checking is a bit more specific than other resources */
|
|
if (rs->dependency->specifier == IA64_RS_AR_UNAT
|
|
&& specs[count].mem_offset.hint
|
|
&& rs->mem_offset.hint)
|
|
{
|
|
if (rs->mem_offset.base == specs[count].mem_offset.base)
|
|
{
|
|
if (((rs->mem_offset.offset >> 3) & 0x3F) ==
|
|
((specs[count].mem_offset.offset >> 3) & 0x3F))
|
|
return 1;
|
|
else
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/* Skip apparent PR write conflicts where both writes are an AND or both
|
|
writes are an OR. */
|
|
if (rs->dependency->specifier == IA64_RS_PR
|
|
|| rs->dependency->specifier == IA64_RS_PRr
|
|
|| rs->dependency->specifier == IA64_RS_PR63)
|
|
{
|
|
if (specs[count].cmp_type != CMP_NONE
|
|
&& specs[count].cmp_type == rs->cmp_type)
|
|
{
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " %s on parallel compare allowed (PR%d)\n",
|
|
dv_mode[rs->dependency->mode],
|
|
rs->dependency->specifier != IA64_RS_PR63 ?
|
|
specs[count].index : 63);
|
|
continue;
|
|
}
|
|
if (md.debug_dv)
|
|
fprintf (stderr,
|
|
" %s on parallel compare conflict %s vs %s on PR%d\n",
|
|
dv_mode[rs->dependency->mode],
|
|
dv_cmp_type[rs->cmp_type],
|
|
dv_cmp_type[specs[count].cmp_type],
|
|
rs->dependency->specifier != IA64_RS_PR63 ?
|
|
specs[count].index : 63);
|
|
|
|
}
|
|
|
|
/* If either resource is not specific, conservatively assume a conflict
|
|
*/
|
|
if (!specs[count].specific || !rs->specific)
|
|
return 2;
|
|
else if (specs[count].index == rs->index)
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Indicate an instruction group break; if INSERT_STOP is non-zero, then
|
|
insert a stop to create the break. Update all resource dependencies
|
|
appropriately. If QP_REGNO is non-zero, only apply the break to resources
|
|
which use the same QP_REGNO and have the link_to_qp_branch flag set.
|
|
If SAVE_CURRENT is non-zero, don't affect resources marked by the current
|
|
instruction. */
|
|
|
|
static void
|
|
insn_group_break (int insert_stop, int qp_regno, int save_current)
|
|
{
|
|
int i;
|
|
|
|
if (insert_stop && md.num_slots_in_use > 0)
|
|
PREV_SLOT.end_of_insn_group = 1;
|
|
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " Insn group break%s",
|
|
(insert_stop ? " (w/stop)" : ""));
|
|
if (qp_regno != 0)
|
|
fprintf (stderr, " effective for QP=%d", qp_regno);
|
|
fprintf (stderr, "\n");
|
|
}
|
|
|
|
i = 0;
|
|
while (i < regdepslen)
|
|
{
|
|
const struct ia64_dependency *dep = regdeps[i].dependency;
|
|
|
|
if (qp_regno != 0
|
|
&& regdeps[i].qp_regno != qp_regno)
|
|
{
|
|
++i;
|
|
continue;
|
|
}
|
|
|
|
if (save_current
|
|
&& CURR_SLOT.src_file == regdeps[i].file
|
|
&& CURR_SLOT.src_line == regdeps[i].line)
|
|
{
|
|
++i;
|
|
continue;
|
|
}
|
|
|
|
/* clear dependencies which are automatically cleared by a stop, or
|
|
those that have reached the appropriate state of insn serialization */
|
|
if (dep->semantics == IA64_DVS_IMPLIED
|
|
|| dep->semantics == IA64_DVS_IMPLIEDF
|
|
|| regdeps[i].insn_srlz == STATE_SRLZ)
|
|
{
|
|
print_dependency ("Removing", i);
|
|
regdeps[i] = regdeps[--regdepslen];
|
|
}
|
|
else
|
|
{
|
|
if (dep->semantics == IA64_DVS_DATA
|
|
|| dep->semantics == IA64_DVS_INSTR
|
|
|| dep->semantics == IA64_DVS_SPECIFIC)
|
|
{
|
|
if (regdeps[i].insn_srlz == STATE_NONE)
|
|
regdeps[i].insn_srlz = STATE_STOP;
|
|
if (regdeps[i].data_srlz == STATE_NONE)
|
|
regdeps[i].data_srlz = STATE_STOP;
|
|
}
|
|
++i;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Add the given resource usage spec to the list of active dependencies. */
|
|
|
|
static void
|
|
mark_resource (struct ia64_opcode *idesc ATTRIBUTE_UNUSED,
|
|
const struct ia64_dependency *dep ATTRIBUTE_UNUSED,
|
|
struct rsrc *spec,
|
|
int depind,
|
|
int path)
|
|
{
|
|
if (regdepslen == regdepstotlen)
|
|
{
|
|
regdepstotlen += 20;
|
|
regdeps = XRESIZEVEC (struct rsrc, regdeps, regdepstotlen);
|
|
}
|
|
|
|
regdeps[regdepslen] = *spec;
|
|
regdeps[regdepslen].depind = depind;
|
|
regdeps[regdepslen].path = path;
|
|
regdeps[regdepslen].file = CURR_SLOT.src_file;
|
|
regdeps[regdepslen].line = CURR_SLOT.src_line;
|
|
|
|
print_dependency ("Adding", regdepslen);
|
|
|
|
++regdepslen;
|
|
}
|
|
|
|
static void
|
|
print_dependency (const char *action, int depind)
|
|
{
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, " %s %s '%s'",
|
|
action, dv_mode[(regdeps[depind].dependency)->mode],
|
|
(regdeps[depind].dependency)->name);
|
|
if (regdeps[depind].specific && regdeps[depind].index >= 0)
|
|
fprintf (stderr, " (%d)", regdeps[depind].index);
|
|
if (regdeps[depind].mem_offset.hint)
|
|
{
|
|
fputs (" ", stderr);
|
|
fprintf_vma (stderr, regdeps[depind].mem_offset.base);
|
|
fputs ("+", stderr);
|
|
fprintf_vma (stderr, regdeps[depind].mem_offset.offset);
|
|
}
|
|
fprintf (stderr, "\n");
|
|
}
|
|
}
|
|
|
|
static void
|
|
instruction_serialization (void)
|
|
{
|
|
int i;
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Instruction serialization\n");
|
|
for (i = 0; i < regdepslen; i++)
|
|
if (regdeps[i].insn_srlz == STATE_STOP)
|
|
regdeps[i].insn_srlz = STATE_SRLZ;
|
|
}
|
|
|
|
static void
|
|
data_serialization (void)
|
|
{
|
|
int i = 0;
|
|
if (md.debug_dv)
|
|
fprintf (stderr, " Data serialization\n");
|
|
while (i < regdepslen)
|
|
{
|
|
if (regdeps[i].data_srlz == STATE_STOP
|
|
/* Note: as of 991210, all "other" dependencies are cleared by a
|
|
data serialization. This might change with new tables */
|
|
|| (regdeps[i].dependency)->semantics == IA64_DVS_OTHER)
|
|
{
|
|
print_dependency ("Removing", i);
|
|
regdeps[i] = regdeps[--regdepslen];
|
|
}
|
|
else
|
|
++i;
|
|
}
|
|
}
|
|
|
|
/* Insert stops and serializations as needed to avoid DVs. */
|
|
|
|
static void
|
|
remove_marked_resource (struct rsrc *rs)
|
|
{
|
|
switch (rs->dependency->semantics)
|
|
{
|
|
case IA64_DVS_SPECIFIC:
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Implementation-specific, assume worst case...\n");
|
|
/* Fall through. */
|
|
case IA64_DVS_INSTR:
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Inserting instr serialization\n");
|
|
if (rs->insn_srlz < STATE_STOP)
|
|
insn_group_break (1, 0, 0);
|
|
if (rs->insn_srlz < STATE_SRLZ)
|
|
{
|
|
struct slot oldslot = CURR_SLOT;
|
|
/* Manually jam a srlz.i insn into the stream */
|
|
memset (&CURR_SLOT, 0, sizeof (CURR_SLOT));
|
|
CURR_SLOT.user_template = -1;
|
|
CURR_SLOT.idesc = ia64_find_opcode ("srlz.i");
|
|
instruction_serialization ();
|
|
md.curr_slot = (md.curr_slot + 1) % NUM_SLOTS;
|
|
if (++md.num_slots_in_use >= NUM_SLOTS)
|
|
emit_one_bundle ();
|
|
CURR_SLOT = oldslot;
|
|
}
|
|
insn_group_break (1, 0, 0);
|
|
break;
|
|
case IA64_DVS_OTHER: /* as of rev2 (991220) of the DV tables, all
|
|
"other" types of DV are eliminated
|
|
by a data serialization */
|
|
case IA64_DVS_DATA:
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Inserting data serialization\n");
|
|
if (rs->data_srlz < STATE_STOP)
|
|
insn_group_break (1, 0, 0);
|
|
{
|
|
struct slot oldslot = CURR_SLOT;
|
|
/* Manually jam a srlz.d insn into the stream */
|
|
memset (&CURR_SLOT, 0, sizeof (CURR_SLOT));
|
|
CURR_SLOT.user_template = -1;
|
|
CURR_SLOT.idesc = ia64_find_opcode ("srlz.d");
|
|
data_serialization ();
|
|
md.curr_slot = (md.curr_slot + 1) % NUM_SLOTS;
|
|
if (++md.num_slots_in_use >= NUM_SLOTS)
|
|
emit_one_bundle ();
|
|
CURR_SLOT = oldslot;
|
|
}
|
|
break;
|
|
case IA64_DVS_IMPLIED:
|
|
case IA64_DVS_IMPLIEDF:
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Inserting stop\n");
|
|
insn_group_break (1, 0, 0);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Check the resources used by the given opcode against the current dependency
|
|
list.
|
|
|
|
The check is run once for each execution path encountered. In this case,
|
|
a unique execution path is the sequence of instructions following a code
|
|
entry point, e.g. the following has three execution paths, one starting
|
|
at L0, one at L1, and one at L2.
|
|
|
|
L0: nop
|
|
L1: add
|
|
L2: add
|
|
br.ret
|
|
*/
|
|
|
|
static void
|
|
check_dependencies (struct ia64_opcode *idesc)
|
|
{
|
|
const struct ia64_opcode_dependency *opdeps = idesc->dependencies;
|
|
int path;
|
|
int i;
|
|
|
|
/* Note that the number of marked resources may change within the
|
|
loop if in auto mode. */
|
|
i = 0;
|
|
while (i < regdepslen)
|
|
{
|
|
struct rsrc *rs = ®deps[i];
|
|
const struct ia64_dependency *dep = rs->dependency;
|
|
int chkind;
|
|
int note;
|
|
int start_over = 0;
|
|
|
|
if (dep->semantics == IA64_DVS_NONE
|
|
|| (chkind = depends_on (rs->depind, idesc)) == -1)
|
|
{
|
|
++i;
|
|
continue;
|
|
}
|
|
|
|
note = NOTE (opdeps->chks[chkind]);
|
|
|
|
/* Check this resource against each execution path seen thus far. */
|
|
for (path = 0; path <= md.path; path++)
|
|
{
|
|
int matchtype;
|
|
|
|
/* If the dependency wasn't on the path being checked, ignore it. */
|
|
if (rs->path < path)
|
|
continue;
|
|
|
|
/* If the QP for this insn implies a QP which has branched, don't
|
|
bother checking. Ed. NOTE: I don't think this check is terribly
|
|
useful; what's the point of generating code which will only be
|
|
reached if its QP is zero?
|
|
This code was specifically inserted to handle the following code,
|
|
based on notes from Intel's DV checking code, where p1 implies p2.
|
|
|
|
mov r4 = 2
|
|
(p2) br.cond L
|
|
(p1) mov r4 = 7
|
|
*/
|
|
if (CURR_SLOT.qp_regno != 0)
|
|
{
|
|
int skip = 0;
|
|
int implies;
|
|
for (implies = 0; implies < qp_implieslen; implies++)
|
|
{
|
|
if (qp_implies[implies].path >= path
|
|
&& qp_implies[implies].p1 == CURR_SLOT.qp_regno
|
|
&& qp_implies[implies].p2_branched)
|
|
{
|
|
skip = 1;
|
|
break;
|
|
}
|
|
}
|
|
if (skip)
|
|
continue;
|
|
}
|
|
|
|
if ((matchtype = resources_match (rs, idesc, note,
|
|
CURR_SLOT.qp_regno, path)) != 0)
|
|
{
|
|
char msg[1024];
|
|
char pathmsg[256] = "";
|
|
char indexmsg[256] = "";
|
|
int certain = (matchtype == 1 && CURR_SLOT.qp_regno == 0);
|
|
|
|
if (path != 0)
|
|
snprintf (pathmsg, sizeof (pathmsg),
|
|
" when entry is at label '%s'",
|
|
md.entry_labels[path - 1]);
|
|
if (matchtype == 1 && rs->index >= 0)
|
|
snprintf (indexmsg, sizeof (indexmsg),
|
|
", specific resource number is %d",
|
|
rs->index);
|
|
snprintf (msg, sizeof (msg),
|
|
"Use of '%s' %s %s dependency '%s' (%s)%s%s",
|
|
idesc->name,
|
|
(certain ? "violates" : "may violate"),
|
|
dv_mode[dep->mode], dep->name,
|
|
dv_sem[dep->semantics],
|
|
pathmsg, indexmsg);
|
|
|
|
if (md.explicit_mode)
|
|
{
|
|
as_warn ("%s", msg);
|
|
if (path < md.path)
|
|
as_warn (_("Only the first path encountering the conflict is reported"));
|
|
as_warn_where (rs->file, rs->line,
|
|
_("This is the location of the conflicting usage"));
|
|
/* Don't bother checking other paths, to avoid duplicating
|
|
the same warning */
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "%s @ %s:%d\n", msg, rs->file, rs->line);
|
|
|
|
remove_marked_resource (rs);
|
|
|
|
/* since the set of dependencies has changed, start over */
|
|
/* FIXME -- since we're removing dvs as we go, we
|
|
probably don't really need to start over... */
|
|
start_over = 1;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (start_over)
|
|
i = 0;
|
|
else
|
|
++i;
|
|
}
|
|
}
|
|
|
|
/* Register new dependencies based on the given opcode. */
|
|
|
|
static void
|
|
mark_resources (struct ia64_opcode *idesc)
|
|
{
|
|
int i;
|
|
const struct ia64_opcode_dependency *opdeps = idesc->dependencies;
|
|
int add_only_qp_reads = 0;
|
|
|
|
/* A conditional branch only uses its resources if it is taken; if it is
|
|
taken, we stop following that path. The other branch types effectively
|
|
*always* write their resources. If it's not taken, register only QP
|
|
reads. */
|
|
if (is_conditional_branch (idesc) || is_interruption_or_rfi (idesc))
|
|
{
|
|
add_only_qp_reads = 1;
|
|
}
|
|
|
|
if (md.debug_dv)
|
|
fprintf (stderr, "Registering '%s' resource usage\n", idesc->name);
|
|
|
|
for (i = 0; i < opdeps->nregs; i++)
|
|
{
|
|
const struct ia64_dependency *dep;
|
|
struct rsrc specs[MAX_SPECS];
|
|
int note;
|
|
int path;
|
|
int count;
|
|
|
|
dep = ia64_find_dependency (opdeps->regs[i]);
|
|
note = NOTE (opdeps->regs[i]);
|
|
|
|
if (add_only_qp_reads
|
|
&& !(dep->mode == IA64_DV_WAR
|
|
&& (dep->specifier == IA64_RS_PR
|
|
|| dep->specifier == IA64_RS_PRr
|
|
|| dep->specifier == IA64_RS_PR63)))
|
|
continue;
|
|
|
|
count = specify_resource (dep, idesc, DV_REG, specs, note, md.path);
|
|
|
|
while (count-- > 0)
|
|
{
|
|
mark_resource (idesc, dep, &specs[count],
|
|
DEP (opdeps->regs[i]), md.path);
|
|
}
|
|
|
|
/* The execution path may affect register values, which may in turn
|
|
affect which indirect-access resources are accessed. */
|
|
switch (dep->specifier)
|
|
{
|
|
default:
|
|
break;
|
|
case IA64_RS_CPUID:
|
|
case IA64_RS_DBR:
|
|
case IA64_RS_IBR:
|
|
case IA64_RS_MSR:
|
|
case IA64_RS_PKR:
|
|
case IA64_RS_PMC:
|
|
case IA64_RS_PMD:
|
|
case IA64_RS_RR:
|
|
for (path = 0; path < md.path; path++)
|
|
{
|
|
count = specify_resource (dep, idesc, DV_REG, specs, note, path);
|
|
while (count-- > 0)
|
|
mark_resource (idesc, dep, &specs[count],
|
|
DEP (opdeps->regs[i]), path);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Remove dependencies when they no longer apply. */
|
|
|
|
static void
|
|
update_dependencies (struct ia64_opcode *idesc)
|
|
{
|
|
int i;
|
|
|
|
if (strcmp (idesc->name, "srlz.i") == 0)
|
|
{
|
|
instruction_serialization ();
|
|
}
|
|
else if (strcmp (idesc->name, "srlz.d") == 0)
|
|
{
|
|
data_serialization ();
|
|
}
|
|
else if (is_interruption_or_rfi (idesc)
|
|
|| is_taken_branch (idesc))
|
|
{
|
|
/* Although technically the taken branch doesn't clear dependencies
|
|
which require a srlz.[id], we don't follow the branch; the next
|
|
instruction is assumed to start with a clean slate. */
|
|
regdepslen = 0;
|
|
md.path = 0;
|
|
}
|
|
else if (is_conditional_branch (idesc)
|
|
&& CURR_SLOT.qp_regno != 0)
|
|
{
|
|
int is_call = strstr (idesc->name, ".call") != NULL;
|
|
|
|
for (i = 0; i < qp_implieslen; i++)
|
|
{
|
|
/* If the conditional branch's predicate is implied by the predicate
|
|
in an existing dependency, remove that dependency. */
|
|
if (qp_implies[i].p2 == CURR_SLOT.qp_regno)
|
|
{
|
|
int depind = 0;
|
|
/* Note that this implied predicate takes a branch so that if
|
|
a later insn generates a DV but its predicate implies this
|
|
one, we can avoid the false DV warning. */
|
|
qp_implies[i].p2_branched = 1;
|
|
while (depind < regdepslen)
|
|
{
|
|
if (regdeps[depind].qp_regno == qp_implies[i].p1)
|
|
{
|
|
print_dependency ("Removing", depind);
|
|
regdeps[depind] = regdeps[--regdepslen];
|
|
}
|
|
else
|
|
++depind;
|
|
}
|
|
}
|
|
}
|
|
/* Any marked resources which have this same predicate should be
|
|
cleared, provided that the QP hasn't been modified between the
|
|
marking instruction and the branch. */
|
|
if (is_call)
|
|
{
|
|
insn_group_break (0, CURR_SLOT.qp_regno, 1);
|
|
}
|
|
else
|
|
{
|
|
i = 0;
|
|
while (i < regdepslen)
|
|
{
|
|
if (regdeps[i].qp_regno == CURR_SLOT.qp_regno
|
|
&& regdeps[i].link_to_qp_branch
|
|
&& (regdeps[i].file != CURR_SLOT.src_file
|
|
|| regdeps[i].line != CURR_SLOT.src_line))
|
|
{
|
|
/* Treat like a taken branch */
|
|
print_dependency ("Removing", i);
|
|
regdeps[i] = regdeps[--regdepslen];
|
|
}
|
|
else
|
|
++i;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Examine the current instruction for dependency violations. */
|
|
|
|
static int
|
|
check_dv (struct ia64_opcode *idesc)
|
|
{
|
|
if (md.debug_dv)
|
|
{
|
|
fprintf (stderr, "Checking %s for violations (line %d, %d/%d)\n",
|
|
idesc->name, CURR_SLOT.src_line,
|
|
idesc->dependencies->nchks,
|
|
idesc->dependencies->nregs);
|
|
}
|
|
|
|
/* Look through the list of currently marked resources; if the current
|
|
instruction has the dependency in its chks list which uses that resource,
|
|
check against the specific resources used. */
|
|
check_dependencies (idesc);
|
|
|
|
/* Look up the instruction's regdeps (RAW writes, WAW writes, and WAR reads),
|
|
then add them to the list of marked resources. */
|
|
mark_resources (idesc);
|
|
|
|
/* There are several types of dependency semantics, and each has its own
|
|
requirements for being cleared
|
|
|
|
Instruction serialization (insns separated by interruption, rfi, or
|
|
writer + srlz.i + reader, all in separate groups) clears DVS_INSTR.
|
|
|
|
Data serialization (instruction serialization, or writer + srlz.d +
|
|
reader, where writer and srlz.d are in separate groups) clears
|
|
DVS_DATA. (This also clears DVS_OTHER, but that is not guaranteed to
|
|
always be the case).
|
|
|
|
Instruction group break (groups separated by stop, taken branch,
|
|
interruption or rfi) clears DVS_IMPLIED and DVS_IMPLIEDF.
|
|
*/
|
|
update_dependencies (idesc);
|
|
|
|
/* Sometimes, knowing a register value allows us to avoid giving a false DV
|
|
warning. Keep track of as many as possible that are useful. */
|
|
note_register_values (idesc);
|
|
|
|
/* We don't need or want this anymore. */
|
|
md.mem_offset.hint = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Translate one line of assembly. Pseudo ops and labels do not show
|
|
here. */
|
|
void
|
|
md_assemble (char *str)
|
|
{
|
|
char *saved_input_line_pointer, *temp;
|
|
const char *mnemonic;
|
|
const struct pseudo_opcode *pdesc;
|
|
struct ia64_opcode *idesc;
|
|
unsigned char qp_regno;
|
|
unsigned int flags;
|
|
int ch;
|
|
|
|
saved_input_line_pointer = input_line_pointer;
|
|
input_line_pointer = str;
|
|
|
|
/* extract the opcode (mnemonic): */
|
|
|
|
ch = get_symbol_name (&temp);
|
|
mnemonic = temp;
|
|
pdesc = (struct pseudo_opcode *) hash_find (md.pseudo_hash, mnemonic);
|
|
if (pdesc)
|
|
{
|
|
(void) restore_line_pointer (ch);
|
|
(*pdesc->handler) (pdesc->arg);
|
|
goto done;
|
|
}
|
|
|
|
/* Find the instruction descriptor matching the arguments. */
|
|
|
|
idesc = ia64_find_opcode (mnemonic);
|
|
(void) restore_line_pointer (ch);
|
|
if (!idesc)
|
|
{
|
|
as_bad (_("Unknown opcode `%s'"), mnemonic);
|
|
goto done;
|
|
}
|
|
|
|
idesc = parse_operands (idesc);
|
|
if (!idesc)
|
|
goto done;
|
|
|
|
/* Handle the dynamic ops we can handle now: */
|
|
if (idesc->type == IA64_TYPE_DYN)
|
|
{
|
|
if (strcmp (idesc->name, "add") == 0)
|
|
{
|
|
if (CURR_SLOT.opnd[2].X_op == O_register
|
|
&& CURR_SLOT.opnd[2].X_add_number < 4)
|
|
mnemonic = "addl";
|
|
else
|
|
mnemonic = "adds";
|
|
ia64_free_opcode (idesc);
|
|
idesc = ia64_find_opcode (mnemonic);
|
|
}
|
|
else if (strcmp (idesc->name, "mov") == 0)
|
|
{
|
|
enum ia64_opnd opnd1, opnd2;
|
|
int rop;
|
|
|
|
opnd1 = idesc->operands[0];
|
|
opnd2 = idesc->operands[1];
|
|
if (opnd1 == IA64_OPND_AR3)
|
|
rop = 0;
|
|
else if (opnd2 == IA64_OPND_AR3)
|
|
rop = 1;
|
|
else
|
|
abort ();
|
|
if (CURR_SLOT.opnd[rop].X_op == O_register)
|
|
{
|
|
if (ar_is_only_in_integer_unit (CURR_SLOT.opnd[rop].X_add_number))
|
|
mnemonic = "mov.i";
|
|
else if (ar_is_only_in_memory_unit (CURR_SLOT.opnd[rop].X_add_number))
|
|
mnemonic = "mov.m";
|
|
else
|
|
rop = -1;
|
|
}
|
|
else
|
|
abort ();
|
|
if (rop >= 0)
|
|
{
|
|
ia64_free_opcode (idesc);
|
|
idesc = ia64_find_opcode (mnemonic);
|
|
while (idesc != NULL
|
|
&& (idesc->operands[0] != opnd1
|
|
|| idesc->operands[1] != opnd2))
|
|
idesc = get_next_opcode (idesc);
|
|
}
|
|
}
|
|
}
|
|
else if (strcmp (idesc->name, "mov.i") == 0
|
|
|| strcmp (idesc->name, "mov.m") == 0)
|
|
{
|
|
enum ia64_opnd opnd1, opnd2;
|
|
int rop;
|
|
|
|
opnd1 = idesc->operands[0];
|
|
opnd2 = idesc->operands[1];
|
|
if (opnd1 == IA64_OPND_AR3)
|
|
rop = 0;
|
|
else if (opnd2 == IA64_OPND_AR3)
|
|
rop = 1;
|
|
else
|
|
abort ();
|
|
if (CURR_SLOT.opnd[rop].X_op == O_register)
|
|
{
|
|
char unit = 'a';
|
|
if (ar_is_only_in_integer_unit (CURR_SLOT.opnd[rop].X_add_number))
|
|
unit = 'i';
|
|
else if (ar_is_only_in_memory_unit (CURR_SLOT.opnd[rop].X_add_number))
|
|
unit = 'm';
|
|
if (unit != 'a' && unit != idesc->name [4])
|
|
as_bad (_("AR %d can only be accessed by %c-unit"),
|
|
(int) (CURR_SLOT.opnd[rop].X_add_number - REG_AR),
|
|
TOUPPER (unit));
|
|
}
|
|
}
|
|
else if (strcmp (idesc->name, "hint.b") == 0)
|
|
{
|
|
switch (md.hint_b)
|
|
{
|
|
case hint_b_ok:
|
|
break;
|
|
case hint_b_warning:
|
|
as_warn (_("hint.b may be treated as nop"));
|
|
break;
|
|
case hint_b_error:
|
|
as_bad (_("hint.b shouldn't be used"));
|
|
break;
|
|
}
|
|
}
|
|
|
|
qp_regno = 0;
|
|
if (md.qp.X_op == O_register)
|
|
{
|
|
qp_regno = md.qp.X_add_number - REG_P;
|
|
md.qp.X_op = O_absent;
|
|
}
|
|
|
|
flags = idesc->flags;
|
|
|
|
if ((flags & IA64_OPCODE_FIRST) != 0)
|
|
{
|
|
/* The alignment frag has to end with a stop bit only if the
|
|
next instruction after the alignment directive has to be
|
|
the first instruction in an instruction group. */
|
|
if (align_frag)
|
|
{
|
|
while (align_frag->fr_type != rs_align_code)
|
|
{
|
|
align_frag = align_frag->fr_next;
|
|
if (!align_frag)
|
|
break;
|
|
}
|
|
/* align_frag can be NULL if there are directives in
|
|
between. */
|
|
if (align_frag && align_frag->fr_next == frag_now)
|
|
align_frag->tc_frag_data = 1;
|
|
}
|
|
|
|
insn_group_break (1, 0, 0);
|
|
}
|
|
align_frag = NULL;
|
|
|
|
if ((flags & IA64_OPCODE_NO_PRED) != 0 && qp_regno != 0)
|
|
{
|
|
as_bad (_("`%s' cannot be predicated"), idesc->name);
|
|
goto done;
|
|
}
|
|
|
|
/* Build the instruction. */
|
|
CURR_SLOT.qp_regno = qp_regno;
|
|
CURR_SLOT.idesc = idesc;
|
|
CURR_SLOT.src_file = as_where (&CURR_SLOT.src_line);
|
|
dwarf2_where (&CURR_SLOT.debug_line);
|
|
dwarf2_consume_line_info ();
|
|
|
|
/* Add unwind entries, if there are any. */
|
|
if (unwind.current_entry)
|
|
{
|
|
CURR_SLOT.unwind_record = unwind.current_entry;
|
|
unwind.current_entry = NULL;
|
|
}
|
|
if (unwind.pending_saves)
|
|
{
|
|
if (unwind.pending_saves->next)
|
|
{
|
|
/* Attach the next pending save to the next slot so that its
|
|
slot number will get set correctly. */
|
|
add_unwind_entry (unwind.pending_saves->next, NOT_A_CHAR);
|
|
unwind.pending_saves = &unwind.pending_saves->next->r.record.p;
|
|
}
|
|
else
|
|
unwind.pending_saves = NULL;
|
|
}
|
|
if (unwind.proc_pending.sym && S_IS_DEFINED (unwind.proc_pending.sym))
|
|
unwind.insn = 1;
|
|
|
|
/* Check for dependency violations. */
|
|
if (md.detect_dv)
|
|
check_dv (idesc);
|
|
|
|
md.curr_slot = (md.curr_slot + 1) % NUM_SLOTS;
|
|
if (++md.num_slots_in_use >= NUM_SLOTS)
|
|
emit_one_bundle ();
|
|
|
|
if ((flags & IA64_OPCODE_LAST) != 0)
|
|
insn_group_break (1, 0, 0);
|
|
|
|
md.last_text_seg = now_seg;
|
|
|
|
done:
|
|
input_line_pointer = saved_input_line_pointer;
|
|
}
|
|
|
|
/* Called when symbol NAME cannot be found in the symbol table.
|
|
Should be used for dynamic valued symbols only. */
|
|
|
|
symbolS *
|
|
md_undefined_symbol (char *name ATTRIBUTE_UNUSED)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
/* Called for any expression that can not be recognized. When the
|
|
function is called, `input_line_pointer' will point to the start of
|
|
the expression. */
|
|
|
|
void
|
|
md_operand (expressionS *e)
|
|
{
|
|
switch (*input_line_pointer)
|
|
{
|
|
case '[':
|
|
++input_line_pointer;
|
|
expression_and_evaluate (e);
|
|
if (*input_line_pointer != ']')
|
|
{
|
|
as_bad (_("Closing bracket missing"));
|
|
goto err;
|
|
}
|
|
else
|
|
{
|
|
if (e->X_op != O_register
|
|
|| e->X_add_number < REG_GR
|
|
|| e->X_add_number > REG_GR + 127)
|
|
{
|
|
as_bad (_("Index must be a general register"));
|
|
e->X_add_number = REG_GR;
|
|
}
|
|
|
|
++input_line_pointer;
|
|
e->X_op = O_index;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
return;
|
|
|
|
err:
|
|
ignore_rest_of_line ();
|
|
}
|
|
|
|
/* Return 1 if it's OK to adjust a reloc by replacing the symbol with
|
|
a section symbol plus some offset. For relocs involving @fptr(),
|
|
directives we don't want such adjustments since we need to have the
|
|
original symbol's name in the reloc. */
|
|
int
|
|
ia64_fix_adjustable (fixS *fix)
|
|
{
|
|
/* Prevent all adjustments to global symbols */
|
|
if (S_IS_EXTERNAL (fix->fx_addsy) || S_IS_WEAK (fix->fx_addsy))
|
|
return 0;
|
|
|
|
switch (fix->fx_r_type)
|
|
{
|
|
case BFD_RELOC_IA64_FPTR64I:
|
|
case BFD_RELOC_IA64_FPTR32MSB:
|
|
case BFD_RELOC_IA64_FPTR32LSB:
|
|
case BFD_RELOC_IA64_FPTR64MSB:
|
|
case BFD_RELOC_IA64_FPTR64LSB:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR22:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR64I:
|
|
return 0;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int
|
|
ia64_force_relocation (fixS *fix)
|
|
{
|
|
switch (fix->fx_r_type)
|
|
{
|
|
case BFD_RELOC_IA64_FPTR64I:
|
|
case BFD_RELOC_IA64_FPTR32MSB:
|
|
case BFD_RELOC_IA64_FPTR32LSB:
|
|
case BFD_RELOC_IA64_FPTR64MSB:
|
|
case BFD_RELOC_IA64_FPTR64LSB:
|
|
|
|
case BFD_RELOC_IA64_LTOFF22:
|
|
case BFD_RELOC_IA64_LTOFF64I:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR22:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR64I:
|
|
case BFD_RELOC_IA64_PLTOFF22:
|
|
case BFD_RELOC_IA64_PLTOFF64I:
|
|
case BFD_RELOC_IA64_PLTOFF64MSB:
|
|
case BFD_RELOC_IA64_PLTOFF64LSB:
|
|
|
|
case BFD_RELOC_IA64_LTOFF22X:
|
|
case BFD_RELOC_IA64_LDXMOV:
|
|
return 1;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return generic_force_reloc (fix);
|
|
}
|
|
|
|
/* Decide from what point a pc-relative relocation is relative to,
|
|
relative to the pc-relative fixup. Er, relatively speaking. */
|
|
long
|
|
ia64_pcrel_from_section (fixS *fix, segT sec)
|
|
{
|
|
unsigned long off = fix->fx_frag->fr_address + fix->fx_where;
|
|
|
|
if (bfd_get_section_flags (stdoutput, sec) & SEC_CODE)
|
|
off &= ~0xfUL;
|
|
|
|
return off;
|
|
}
|
|
|
|
|
|
/* Used to emit section-relative relocs for the dwarf2 debug data. */
|
|
void
|
|
ia64_dwarf2_emit_offset (symbolS *symbol, unsigned int size)
|
|
{
|
|
expressionS exp;
|
|
|
|
exp.X_op = O_pseudo_fixup;
|
|
exp.X_op_symbol = pseudo_func[FUNC_SEC_RELATIVE].u.sym;
|
|
exp.X_add_number = 0;
|
|
exp.X_add_symbol = symbol;
|
|
emit_expr (&exp, size);
|
|
}
|
|
|
|
/* This is called whenever some data item (not an instruction) needs a
|
|
fixup. We pick the right reloc code depending on the byteorder
|
|
currently in effect. */
|
|
void
|
|
ia64_cons_fix_new (fragS *f, int where, int nbytes, expressionS *exp,
|
|
bfd_reloc_code_real_type code)
|
|
{
|
|
fixS *fix;
|
|
|
|
switch (nbytes)
|
|
{
|
|
/* There are no reloc for 8 and 16 bit quantities, but we allow
|
|
them here since they will work fine as long as the expression
|
|
is fully defined at the end of the pass over the source file. */
|
|
case 1: code = BFD_RELOC_8; break;
|
|
case 2: code = BFD_RELOC_16; break;
|
|
case 4:
|
|
if (target_big_endian)
|
|
code = BFD_RELOC_IA64_DIR32MSB;
|
|
else
|
|
code = BFD_RELOC_IA64_DIR32LSB;
|
|
break;
|
|
|
|
case 8:
|
|
/* In 32-bit mode, data8 could mean function descriptors too. */
|
|
if (exp->X_op == O_pseudo_fixup
|
|
&& exp->X_op_symbol
|
|
&& S_GET_VALUE (exp->X_op_symbol) == FUNC_IPLT_RELOC
|
|
&& !(md.flags & EF_IA_64_ABI64))
|
|
{
|
|
if (target_big_endian)
|
|
code = BFD_RELOC_IA64_IPLTMSB;
|
|
else
|
|
code = BFD_RELOC_IA64_IPLTLSB;
|
|
exp->X_op = O_symbol;
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
if (target_big_endian)
|
|
code = BFD_RELOC_IA64_DIR64MSB;
|
|
else
|
|
code = BFD_RELOC_IA64_DIR64LSB;
|
|
break;
|
|
}
|
|
|
|
case 16:
|
|
if (exp->X_op == O_pseudo_fixup
|
|
&& exp->X_op_symbol
|
|
&& S_GET_VALUE (exp->X_op_symbol) == FUNC_IPLT_RELOC)
|
|
{
|
|
if (target_big_endian)
|
|
code = BFD_RELOC_IA64_IPLTMSB;
|
|
else
|
|
code = BFD_RELOC_IA64_IPLTLSB;
|
|
exp->X_op = O_symbol;
|
|
break;
|
|
}
|
|
/* FALLTHRU */
|
|
|
|
default:
|
|
as_bad (_("Unsupported fixup size %d"), nbytes);
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
if (exp->X_op == O_pseudo_fixup)
|
|
{
|
|
exp->X_op = O_symbol;
|
|
code = ia64_gen_real_reloc_type (exp->X_op_symbol, code);
|
|
/* ??? If code unchanged, unsupported. */
|
|
}
|
|
|
|
fix = fix_new_exp (f, where, nbytes, exp, 0, code);
|
|
/* We need to store the byte order in effect in case we're going
|
|
to fix an 8 or 16 bit relocation (for which there no real
|
|
relocs available). See md_apply_fix(). */
|
|
fix->tc_fix_data.bigendian = target_big_endian;
|
|
}
|
|
|
|
/* Return the actual relocation we wish to associate with the pseudo
|
|
reloc described by SYM and R_TYPE. SYM should be one of the
|
|
symbols in the pseudo_func array, or NULL. */
|
|
|
|
static bfd_reloc_code_real_type
|
|
ia64_gen_real_reloc_type (struct symbol *sym, bfd_reloc_code_real_type r_type)
|
|
{
|
|
bfd_reloc_code_real_type newr = 0;
|
|
const char *type = NULL, *suffix = "";
|
|
|
|
if (sym == NULL)
|
|
{
|
|
return r_type;
|
|
}
|
|
|
|
switch (S_GET_VALUE (sym))
|
|
{
|
|
case FUNC_FPTR_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_FPTR64I; break;
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_FPTR32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_FPTR32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_FPTR64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_FPTR64LSB; break;
|
|
default: type = "FPTR"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_GP_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_GPREL22; break;
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_GPREL64I; break;
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_GPREL32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_GPREL32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_GPREL64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_GPREL64LSB; break;
|
|
default: type = "GPREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_LTOFF22; break;
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_LTOFF64I; break;
|
|
default: type = "LTOFF"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_RELATIVE_X:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_LTOFF22X; break;
|
|
default: type = "LTOFF"; suffix = "X"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_PC_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_PCREL22; break;
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_PCREL64I; break;
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_PCREL32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_PCREL32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_PCREL64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_PCREL64LSB; break;
|
|
default: type = "PCREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_PLT_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_PLTOFF22; break;
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_PLTOFF64I; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_PLTOFF64MSB;break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_PLTOFF64LSB;break;
|
|
default: type = "PLTOFF"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_SEC_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_SECREL32MSB;break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_SECREL32LSB;break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_SECREL64MSB;break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_SECREL64LSB;break;
|
|
default: type = "SECREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_SEG_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_SEGREL32MSB;break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_SEGREL32LSB;break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_SEGREL64MSB;break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_SEGREL64LSB;break;
|
|
default: type = "SEGREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LTV_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR32MSB: newr = BFD_RELOC_IA64_LTV32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB: newr = BFD_RELOC_IA64_LTV32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_LTV64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_LTV64LSB; break;
|
|
default: type = "LTV"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_FPTR_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR22; break;
|
|
case BFD_RELOC_IA64_IMM64:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR64I; break;
|
|
case BFD_RELOC_IA64_DIR32MSB:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB:
|
|
newr = BFD_RELOC_IA64_LTOFF_FPTR64LSB; break;
|
|
default:
|
|
type = "LTOFF_FPTR"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_TP_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM14: newr = BFD_RELOC_IA64_TPREL14; break;
|
|
case BFD_RELOC_IA64_IMM22: newr = BFD_RELOC_IA64_TPREL22; break;
|
|
case BFD_RELOC_IA64_IMM64: newr = BFD_RELOC_IA64_TPREL64I; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: newr = BFD_RELOC_IA64_TPREL64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: newr = BFD_RELOC_IA64_TPREL64LSB; break;
|
|
default: type = "TPREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_TP_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22:
|
|
newr = BFD_RELOC_IA64_LTOFF_TPREL22; break;
|
|
default:
|
|
type = "LTOFF_TPREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_DTP_MODULE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR64MSB:
|
|
newr = BFD_RELOC_IA64_DTPMOD64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB:
|
|
newr = BFD_RELOC_IA64_DTPMOD64LSB; break;
|
|
default:
|
|
type = "DTPMOD"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_DTP_MODULE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22:
|
|
newr = BFD_RELOC_IA64_LTOFF_DTPMOD22; break;
|
|
default:
|
|
type = "LTOFF_DTPMOD"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_DTP_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR32MSB:
|
|
newr = BFD_RELOC_IA64_DTPREL32MSB; break;
|
|
case BFD_RELOC_IA64_DIR32LSB:
|
|
newr = BFD_RELOC_IA64_DTPREL32LSB; break;
|
|
case BFD_RELOC_IA64_DIR64MSB:
|
|
newr = BFD_RELOC_IA64_DTPREL64MSB; break;
|
|
case BFD_RELOC_IA64_DIR64LSB:
|
|
newr = BFD_RELOC_IA64_DTPREL64LSB; break;
|
|
case BFD_RELOC_IA64_IMM14:
|
|
newr = BFD_RELOC_IA64_DTPREL14; break;
|
|
case BFD_RELOC_IA64_IMM22:
|
|
newr = BFD_RELOC_IA64_DTPREL22; break;
|
|
case BFD_RELOC_IA64_IMM64:
|
|
newr = BFD_RELOC_IA64_DTPREL64I; break;
|
|
default:
|
|
type = "DTPREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_LT_DTP_RELATIVE:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IMM22:
|
|
newr = BFD_RELOC_IA64_LTOFF_DTPREL22; break;
|
|
default:
|
|
type = "LTOFF_DTPREL"; break;
|
|
}
|
|
break;
|
|
|
|
case FUNC_IPLT_RELOC:
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_IPLTMSB: return r_type;
|
|
case BFD_RELOC_IA64_IPLTLSB: return r_type;
|
|
default: type = "IPLT"; break;
|
|
}
|
|
break;
|
|
|
|
#ifdef TE_VMS
|
|
case FUNC_SLOTCOUNT_RELOC:
|
|
return DUMMY_RELOC_IA64_SLOTCOUNT;
|
|
#endif
|
|
|
|
default:
|
|
abort ();
|
|
}
|
|
|
|
if (newr)
|
|
return newr;
|
|
else
|
|
{
|
|
int width;
|
|
|
|
if (!type)
|
|
abort ();
|
|
switch (r_type)
|
|
{
|
|
case BFD_RELOC_IA64_DIR32MSB: width = 32; suffix = "MSB"; break;
|
|
case BFD_RELOC_IA64_DIR32LSB: width = 32; suffix = "LSB"; break;
|
|
case BFD_RELOC_IA64_DIR64MSB: width = 64; suffix = "MSB"; break;
|
|
case BFD_RELOC_IA64_DIR64LSB: width = 64; suffix = "LSB"; break;
|
|
case BFD_RELOC_UNUSED: width = 13; break;
|
|
case BFD_RELOC_IA64_IMM14: width = 14; break;
|
|
case BFD_RELOC_IA64_IMM22: width = 22; break;
|
|
case BFD_RELOC_IA64_IMM64: width = 64; suffix = "I"; break;
|
|
default: abort ();
|
|
}
|
|
|
|
/* This should be an error, but since previously there wasn't any
|
|
diagnostic here, don't make it fail because of this for now. */
|
|
as_warn (_("Cannot express %s%d%s relocation"), type, width, suffix);
|
|
return r_type;
|
|
}
|
|
}
|
|
|
|
/* Here is where generate the appropriate reloc for pseudo relocation
|
|
functions. */
|
|
void
|
|
ia64_validate_fix (fixS *fix)
|
|
{
|
|
switch (fix->fx_r_type)
|
|
{
|
|
case BFD_RELOC_IA64_FPTR64I:
|
|
case BFD_RELOC_IA64_FPTR32MSB:
|
|
case BFD_RELOC_IA64_FPTR64LSB:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR22:
|
|
case BFD_RELOC_IA64_LTOFF_FPTR64I:
|
|
if (fix->fx_offset != 0)
|
|
as_bad_where (fix->fx_file, fix->fx_line,
|
|
_("No addend allowed in @fptr() relocation"));
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
fix_insn (fixS *fix, const struct ia64_operand *odesc, valueT value)
|
|
{
|
|
bfd_vma insn[3], t0, t1, control_bits;
|
|
const char *err;
|
|
char *fixpos;
|
|
long slot;
|
|
|
|
slot = fix->fx_where & 0x3;
|
|
fixpos = fix->fx_frag->fr_literal + (fix->fx_where - slot);
|
|
|
|
/* Bundles are always in little-endian byte order */
|
|
t0 = bfd_getl64 (fixpos);
|
|
t1 = bfd_getl64 (fixpos + 8);
|
|
control_bits = t0 & 0x1f;
|
|
insn[0] = (t0 >> 5) & 0x1ffffffffffLL;
|
|
insn[1] = ((t0 >> 46) & 0x3ffff) | ((t1 & 0x7fffff) << 18);
|
|
insn[2] = (t1 >> 23) & 0x1ffffffffffLL;
|
|
|
|
err = NULL;
|
|
if (odesc - elf64_ia64_operands == IA64_OPND_IMMU64)
|
|
{
|
|
insn[1] = (value >> 22) & 0x1ffffffffffLL;
|
|
insn[2] |= (((value & 0x7f) << 13)
|
|
| (((value >> 7) & 0x1ff) << 27)
|
|
| (((value >> 16) & 0x1f) << 22)
|
|
| (((value >> 21) & 0x1) << 21)
|
|
| (((value >> 63) & 0x1) << 36));
|
|
}
|
|
else if (odesc - elf64_ia64_operands == IA64_OPND_IMMU62)
|
|
{
|
|
if (value & ~0x3fffffffffffffffULL)
|
|
err = _("integer operand out of range");
|
|
insn[1] = (value >> 21) & 0x1ffffffffffLL;
|
|
insn[2] |= (((value & 0xfffff) << 6) | (((value >> 20) & 0x1) << 36));
|
|
}
|
|
else if (odesc - elf64_ia64_operands == IA64_OPND_TGT64)
|
|
{
|
|
value >>= 4;
|
|
insn[1] = ((value >> 20) & 0x7fffffffffLL) << 2;
|
|
insn[2] |= ((((value >> 59) & 0x1) << 36)
|
|
| (((value >> 0) & 0xfffff) << 13));
|
|
}
|
|
else
|
|
err = (*odesc->insert) (odesc, value, insn + slot);
|
|
|
|
if (err)
|
|
as_bad_where (fix->fx_file, fix->fx_line, "%s", err);
|
|
|
|
t0 = control_bits | (insn[0] << 5) | (insn[1] << 46);
|
|
t1 = ((insn[1] >> 18) & 0x7fffff) | (insn[2] << 23);
|
|
number_to_chars_littleendian (fixpos + 0, t0, 8);
|
|
number_to_chars_littleendian (fixpos + 8, t1, 8);
|
|
}
|
|
|
|
/* Attempt to simplify or even eliminate a fixup. The return value is
|
|
ignored; perhaps it was once meaningful, but now it is historical.
|
|
To indicate that a fixup has been eliminated, set FIXP->FX_DONE.
|
|
|
|
If fixp->fx_addsy is non-NULL, we'll have to generate a reloc entry
|
|
(if possible). */
|
|
|
|
void
|
|
md_apply_fix (fixS *fix, valueT *valP, segT seg ATTRIBUTE_UNUSED)
|
|
{
|
|
char *fixpos;
|
|
valueT value = *valP;
|
|
|
|
fixpos = fix->fx_frag->fr_literal + fix->fx_where;
|
|
|
|
if (fix->fx_pcrel)
|
|
{
|
|
switch (fix->fx_r_type)
|
|
{
|
|
case BFD_RELOC_IA64_PCREL21B: break;
|
|
case BFD_RELOC_IA64_PCREL21BI: break;
|
|
case BFD_RELOC_IA64_PCREL21F: break;
|
|
case BFD_RELOC_IA64_PCREL21M: break;
|
|
case BFD_RELOC_IA64_PCREL60B: break;
|
|
case BFD_RELOC_IA64_PCREL22: break;
|
|
case BFD_RELOC_IA64_PCREL64I: break;
|
|
case BFD_RELOC_IA64_PCREL32MSB: break;
|
|
case BFD_RELOC_IA64_PCREL32LSB: break;
|
|
case BFD_RELOC_IA64_PCREL64MSB: break;
|
|
case BFD_RELOC_IA64_PCREL64LSB: break;
|
|
default:
|
|
fix->fx_r_type = ia64_gen_real_reloc_type (pseudo_func[FUNC_PC_RELATIVE].u.sym,
|
|
fix->fx_r_type);
|
|
break;
|
|
}
|
|
}
|
|
if (fix->fx_addsy)
|
|
{
|
|
switch ((unsigned) fix->fx_r_type)
|
|
{
|
|
case BFD_RELOC_UNUSED:
|
|
/* This must be a TAG13 or TAG13b operand. There are no external
|
|
relocs defined for them, so we must give an error. */
|
|
as_bad_where (fix->fx_file, fix->fx_line,
|
|
_("%s must have a constant value"),
|
|
elf64_ia64_operands[fix->tc_fix_data.opnd].desc);
|
|
fix->fx_done = 1;
|
|
return;
|
|
|
|
case BFD_RELOC_IA64_TPREL14:
|
|
case BFD_RELOC_IA64_TPREL22:
|
|
case BFD_RELOC_IA64_TPREL64I:
|
|
case BFD_RELOC_IA64_LTOFF_TPREL22:
|
|
case BFD_RELOC_IA64_LTOFF_DTPMOD22:
|
|
case BFD_RELOC_IA64_DTPREL14:
|
|
case BFD_RELOC_IA64_DTPREL22:
|
|
case BFD_RELOC_IA64_DTPREL64I:
|
|
case BFD_RELOC_IA64_LTOFF_DTPREL22:
|
|
S_SET_THREAD_LOCAL (fix->fx_addsy);
|
|
break;
|
|
|
|
#ifdef TE_VMS
|
|
case DUMMY_RELOC_IA64_SLOTCOUNT:
|
|
as_bad_where (fix->fx_file, fix->fx_line,
|
|
_("cannot resolve @slotcount parameter"));
|
|
fix->fx_done = 1;
|
|
return;
|
|
#endif
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
else if (fix->tc_fix_data.opnd == IA64_OPND_NIL)
|
|
{
|
|
#ifdef TE_VMS
|
|
if (fix->fx_r_type == DUMMY_RELOC_IA64_SLOTCOUNT)
|
|
{
|
|
/* For @slotcount, convert an addresses difference to a slots
|
|
difference. */
|
|
valueT v;
|
|
|
|
v = (value >> 4) * 3;
|
|
switch (value & 0x0f)
|
|
{
|
|
case 0:
|
|
case 1:
|
|
case 2:
|
|
v += value & 0x0f;
|
|
break;
|
|
case 0x0f:
|
|
v += 2;
|
|
break;
|
|
case 0x0e:
|
|
v += 1;
|
|
break;
|
|
default:
|
|
as_bad (_("invalid @slotcount value"));
|
|
}
|
|
value = v;
|
|
}
|
|
#endif
|
|
|
|
if (fix->tc_fix_data.bigendian)
|
|
number_to_chars_bigendian (fixpos, value, fix->fx_size);
|
|
else
|
|
number_to_chars_littleendian (fixpos, value, fix->fx_size);
|
|
fix->fx_done = 1;
|
|
}
|
|
else
|
|
{
|
|
fix_insn (fix, elf64_ia64_operands + fix->tc_fix_data.opnd, value);
|
|
fix->fx_done = 1;
|
|
}
|
|
}
|
|
|
|
/* Generate the BFD reloc to be stuck in the object file from the
|
|
fixup used internally in the assembler. */
|
|
|
|
arelent *
|
|
tc_gen_reloc (asection *sec ATTRIBUTE_UNUSED, fixS *fixp)
|
|
{
|
|
arelent *reloc;
|
|
|
|
reloc = XNEW (arelent);
|
|
reloc->sym_ptr_ptr = XNEW (asymbol *);
|
|
*reloc->sym_ptr_ptr = symbol_get_bfdsym (fixp->fx_addsy);
|
|
reloc->address = fixp->fx_frag->fr_address + fixp->fx_where;
|
|
reloc->addend = fixp->fx_offset;
|
|
reloc->howto = bfd_reloc_type_lookup (stdoutput, fixp->fx_r_type);
|
|
|
|
if (!reloc->howto)
|
|
{
|
|
as_bad_where (fixp->fx_file, fixp->fx_line,
|
|
_("Cannot represent %s relocation in object file"),
|
|
bfd_get_reloc_code_name (fixp->fx_r_type));
|
|
free (reloc);
|
|
return NULL;
|
|
}
|
|
return reloc;
|
|
}
|
|
|
|
/* Turn a string in input_line_pointer into a floating point constant
|
|
of type TYPE, and store the appropriate bytes in *LIT. The number
|
|
of LITTLENUMS emitted is stored in *SIZE. An error message is
|
|
returned, or NULL on OK. */
|
|
|
|
#define MAX_LITTLENUMS 5
|
|
|
|
const char *
|
|
md_atof (int type, char *lit, int *size)
|
|
{
|
|
LITTLENUM_TYPE words[MAX_LITTLENUMS];
|
|
char *t;
|
|
int prec;
|
|
|
|
switch (type)
|
|
{
|
|
/* IEEE floats */
|
|
case 'f':
|
|
case 'F':
|
|
case 's':
|
|
case 'S':
|
|
prec = 2;
|
|
break;
|
|
|
|
case 'd':
|
|
case 'D':
|
|
case 'r':
|
|
case 'R':
|
|
prec = 4;
|
|
break;
|
|
|
|
case 'x':
|
|
case 'X':
|
|
case 'p':
|
|
case 'P':
|
|
prec = 5;
|
|
break;
|
|
|
|
default:
|
|
*size = 0;
|
|
return _("Unrecognized or unsupported floating point constant");
|
|
}
|
|
t = atof_ieee (input_line_pointer, type, words);
|
|
if (t)
|
|
input_line_pointer = t;
|
|
|
|
(*ia64_float_to_chars) (lit, words, prec);
|
|
|
|
if (type == 'X')
|
|
{
|
|
/* It is 10 byte floating point with 6 byte padding. */
|
|
memset (&lit [10], 0, 6);
|
|
*size = 8 * sizeof (LITTLENUM_TYPE);
|
|
}
|
|
else
|
|
*size = prec * sizeof (LITTLENUM_TYPE);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Handle ia64 specific semantics of the align directive. */
|
|
|
|
void
|
|
ia64_md_do_align (int n ATTRIBUTE_UNUSED,
|
|
const char *fill ATTRIBUTE_UNUSED,
|
|
int len ATTRIBUTE_UNUSED,
|
|
int max ATTRIBUTE_UNUSED)
|
|
{
|
|
if (subseg_text_p (now_seg))
|
|
ia64_flush_insns ();
|
|
}
|
|
|
|
/* This is called from HANDLE_ALIGN in write.c. Fill in the contents
|
|
of an rs_align_code fragment. */
|
|
|
|
void
|
|
ia64_handle_align (fragS *fragp)
|
|
{
|
|
int bytes;
|
|
char *p;
|
|
const unsigned char *nop_type;
|
|
|
|
if (fragp->fr_type != rs_align_code)
|
|
return;
|
|
|
|
/* Check if this frag has to end with a stop bit. */
|
|
nop_type = fragp->tc_frag_data ? le_nop_stop : le_nop;
|
|
|
|
bytes = fragp->fr_next->fr_address - fragp->fr_address - fragp->fr_fix;
|
|
p = fragp->fr_literal + fragp->fr_fix;
|
|
|
|
/* If no paddings are needed, we check if we need a stop bit. */
|
|
if (!bytes && fragp->tc_frag_data)
|
|
{
|
|
if (fragp->fr_fix < 16)
|
|
#if 1
|
|
/* FIXME: It won't work with
|
|
.align 16
|
|
alloc r32=ar.pfs,1,2,4,0
|
|
*/
|
|
;
|
|
#else
|
|
as_bad_where (fragp->fr_file, fragp->fr_line,
|
|
_("Can't add stop bit to mark end of instruction group"));
|
|
#endif
|
|
else
|
|
/* Bundles are always in little-endian byte order. Make sure
|
|
the previous bundle has the stop bit. */
|
|
*(p - 16) |= 1;
|
|
}
|
|
|
|
/* Make sure we are on a 16-byte boundary, in case someone has been
|
|
putting data into a text section. */
|
|
if (bytes & 15)
|
|
{
|
|
int fix = bytes & 15;
|
|
memset (p, 0, fix);
|
|
p += fix;
|
|
bytes -= fix;
|
|
fragp->fr_fix += fix;
|
|
}
|
|
|
|
/* Instruction bundles are always little-endian. */
|
|
memcpy (p, nop_type, 16);
|
|
fragp->fr_var = 16;
|
|
}
|
|
|
|
static void
|
|
ia64_float_to_chars_bigendian (char *lit, LITTLENUM_TYPE *words,
|
|
int prec)
|
|
{
|
|
while (prec--)
|
|
{
|
|
number_to_chars_bigendian (lit, (long) (*words++),
|
|
sizeof (LITTLENUM_TYPE));
|
|
lit += sizeof (LITTLENUM_TYPE);
|
|
}
|
|
}
|
|
|
|
static void
|
|
ia64_float_to_chars_littleendian (char *lit, LITTLENUM_TYPE *words,
|
|
int prec)
|
|
{
|
|
while (prec--)
|
|
{
|
|
number_to_chars_littleendian (lit, (long) (words[prec]),
|
|
sizeof (LITTLENUM_TYPE));
|
|
lit += sizeof (LITTLENUM_TYPE);
|
|
}
|
|
}
|
|
|
|
void
|
|
ia64_elf_section_change_hook (void)
|
|
{
|
|
if (elf_section_type (now_seg) == SHT_IA_64_UNWIND
|
|
&& elf_linked_to_section (now_seg) == NULL)
|
|
elf_linked_to_section (now_seg) = text_section;
|
|
dot_byteorder (-1);
|
|
}
|
|
|
|
/* Check if a label should be made global. */
|
|
void
|
|
ia64_check_label (symbolS *label)
|
|
{
|
|
if (*input_line_pointer == ':')
|
|
{
|
|
S_SET_EXTERNAL (label);
|
|
input_line_pointer++;
|
|
}
|
|
}
|
|
|
|
/* Used to remember where .alias and .secalias directives are seen. We
|
|
will rename symbol and section names when we are about to output
|
|
the relocatable file. */
|
|
struct alias
|
|
{
|
|
const char *file; /* The file where the directive is seen. */
|
|
unsigned int line; /* The line number the directive is at. */
|
|
const char *name; /* The original name of the symbol. */
|
|
};
|
|
|
|
/* Called for .alias and .secalias directives. If SECTION is 1, it is
|
|
.secalias. Otherwise, it is .alias. */
|
|
static void
|
|
dot_alias (int section)
|
|
{
|
|
char *name, *alias;
|
|
char delim;
|
|
char *end_name;
|
|
int len;
|
|
const char *error_string;
|
|
struct alias *h;
|
|
const char *a;
|
|
struct hash_control *ahash, *nhash;
|
|
const char *kind;
|
|
|
|
delim = get_symbol_name (&name);
|
|
end_name = input_line_pointer;
|
|
*end_name = delim;
|
|
|
|
if (name == end_name)
|
|
{
|
|
as_bad (_("expected symbol name"));
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
SKIP_WHITESPACE_AFTER_NAME ();
|
|
|
|
if (*input_line_pointer != ',')
|
|
{
|
|
*end_name = 0;
|
|
as_bad (_("expected comma after \"%s\""), name);
|
|
*end_name = delim;
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
input_line_pointer++;
|
|
*end_name = 0;
|
|
ia64_canonicalize_symbol_name (name);
|
|
|
|
/* We call demand_copy_C_string to check if alias string is valid.
|
|
There should be a closing `"' and no `\0' in the string. */
|
|
alias = demand_copy_C_string (&len);
|
|
if (alias == NULL)
|
|
{
|
|
ignore_rest_of_line ();
|
|
return;
|
|
}
|
|
|
|
/* Make a copy of name string. */
|
|
len = strlen (name) + 1;
|
|
obstack_grow (¬es, name, len);
|
|
name = obstack_finish (¬es);
|
|
|
|
if (section)
|
|
{
|
|
kind = "section";
|
|
ahash = secalias_hash;
|
|
nhash = secalias_name_hash;
|
|
}
|
|
else
|
|
{
|
|
kind = "symbol";
|
|
ahash = alias_hash;
|
|
nhash = alias_name_hash;
|
|
}
|
|
|
|
/* Check if alias has been used before. */
|
|
h = (struct alias *) hash_find (ahash, alias);
|
|
if (h)
|
|
{
|
|
if (strcmp (h->name, name))
|
|
as_bad (_("`%s' is already the alias of %s `%s'"),
|
|
alias, kind, h->name);
|
|
goto out;
|
|
}
|
|
|
|
/* Check if name already has an alias. */
|
|
a = (const char *) hash_find (nhash, name);
|
|
if (a)
|
|
{
|
|
if (strcmp (a, alias))
|
|
as_bad (_("%s `%s' already has an alias `%s'"), kind, name, a);
|
|
goto out;
|
|
}
|
|
|
|
h = XNEW (struct alias);
|
|
h->file = as_where (&h->line);
|
|
h->name = name;
|
|
|
|
error_string = hash_jam (ahash, alias, (void *) h);
|
|
if (error_string)
|
|
{
|
|
as_fatal (_("inserting \"%s\" into %s alias hash table failed: %s"),
|
|
alias, kind, error_string);
|
|
goto out;
|
|
}
|
|
|
|
error_string = hash_jam (nhash, name, (void *) alias);
|
|
if (error_string)
|
|
{
|
|
as_fatal (_("inserting \"%s\" into %s name hash table failed: %s"),
|
|
alias, kind, error_string);
|
|
out:
|
|
obstack_free (¬es, name);
|
|
obstack_free (¬es, alias);
|
|
}
|
|
|
|
demand_empty_rest_of_line ();
|
|
}
|
|
|
|
/* It renames the original symbol name to its alias. */
|
|
static void
|
|
do_alias (const char *alias, void *value)
|
|
{
|
|
struct alias *h = (struct alias *) value;
|
|
symbolS *sym = symbol_find (h->name);
|
|
|
|
if (sym == NULL)
|
|
{
|
|
#ifdef TE_VMS
|
|
/* Uses .alias extensively to alias CRTL functions to same with
|
|
decc$ prefix. Sometimes function gets optimized away and a
|
|
warning results, which should be suppressed. */
|
|
if (strncmp (alias, "decc$", 5) != 0)
|
|
#endif
|
|
as_warn_where (h->file, h->line,
|
|
_("symbol `%s' aliased to `%s' is not used"),
|
|
h->name, alias);
|
|
}
|
|
else
|
|
S_SET_NAME (sym, (char *) alias);
|
|
}
|
|
|
|
/* Called from write_object_file. */
|
|
void
|
|
ia64_adjust_symtab (void)
|
|
{
|
|
hash_traverse (alias_hash, do_alias);
|
|
}
|
|
|
|
/* It renames the original section name to its alias. */
|
|
static void
|
|
do_secalias (const char *alias, void *value)
|
|
{
|
|
struct alias *h = (struct alias *) value;
|
|
segT sec = bfd_get_section_by_name (stdoutput, h->name);
|
|
|
|
if (sec == NULL)
|
|
as_warn_where (h->file, h->line,
|
|
_("section `%s' aliased to `%s' is not used"),
|
|
h->name, alias);
|
|
else
|
|
sec->name = alias;
|
|
}
|
|
|
|
/* Called from write_object_file. */
|
|
void
|
|
ia64_frob_file (void)
|
|
{
|
|
hash_traverse (secalias_hash, do_secalias);
|
|
}
|
|
|
|
#ifdef TE_VMS
|
|
#define NT_VMS_MHD 1
|
|
#define NT_VMS_LNM 2
|
|
|
|
/* Integrity VMS 8.x identifies it's ELF modules with a standard ELF
|
|
.note section. */
|
|
|
|
/* Manufacture a VMS-like time string. */
|
|
static void
|
|
get_vms_time (char *Now)
|
|
{
|
|
char *pnt;
|
|
time_t timeb;
|
|
|
|
time (&timeb);
|
|
pnt = ctime (&timeb);
|
|
pnt[3] = 0;
|
|
pnt[7] = 0;
|
|
pnt[10] = 0;
|
|
pnt[16] = 0;
|
|
pnt[24] = 0;
|
|
sprintf (Now, "%2s-%3s-%s %s", pnt + 8, pnt + 4, pnt + 20, pnt + 11);
|
|
}
|
|
|
|
void
|
|
ia64_vms_note (void)
|
|
{
|
|
char *p;
|
|
asection *seg = now_seg;
|
|
subsegT subseg = now_subseg;
|
|
asection *secp = NULL;
|
|
char *bname;
|
|
char buf [256];
|
|
symbolS *sym;
|
|
|
|
/* Create the .note section. */
|
|
|
|
secp = subseg_new (".note", 0);
|
|
bfd_set_section_flags (stdoutput,
|
|
secp,
|
|
SEC_HAS_CONTENTS | SEC_READONLY);
|
|
|
|
/* Module header note (MHD). */
|
|
bname = xstrdup (lbasename (out_file_name));
|
|
if ((p = strrchr (bname, '.')))
|
|
*p = '\0';
|
|
|
|
/* VMS note header is 24 bytes long. */
|
|
p = frag_more (8 + 8 + 8);
|
|
number_to_chars_littleendian (p + 0, 8, 8);
|
|
number_to_chars_littleendian (p + 8, 40 + strlen (bname), 8);
|
|
number_to_chars_littleendian (p + 16, NT_VMS_MHD, 8);
|
|
|
|
p = frag_more (8);
|
|
strcpy (p, "IPF/VMS");
|
|
|
|
p = frag_more (17 + 17 + strlen (bname) + 1 + 5);
|
|
get_vms_time (p);
|
|
strcpy (p + 17, "24-FEB-2005 15:00");
|
|
p += 17 + 17;
|
|
strcpy (p, bname);
|
|
p += strlen (bname) + 1;
|
|
free (bname);
|
|
strcpy (p, "V1.0");
|
|
|
|
frag_align (3, 0, 0);
|
|
|
|
/* Language processor name note. */
|
|
sprintf (buf, "GNU assembler version %s (%s) using BFD version %s",
|
|
VERSION, TARGET_ALIAS, BFD_VERSION_STRING);
|
|
|
|
p = frag_more (8 + 8 + 8);
|
|
number_to_chars_littleendian (p + 0, 8, 8);
|
|
number_to_chars_littleendian (p + 8, strlen (buf) + 1, 8);
|
|
number_to_chars_littleendian (p + 16, NT_VMS_LNM, 8);
|
|
|
|
p = frag_more (8);
|
|
strcpy (p, "IPF/VMS");
|
|
|
|
p = frag_more (strlen (buf) + 1);
|
|
strcpy (p, buf);
|
|
|
|
frag_align (3, 0, 0);
|
|
|
|
secp = subseg_new (".vms_display_name_info", 0);
|
|
bfd_set_section_flags (stdoutput,
|
|
secp,
|
|
SEC_HAS_CONTENTS | SEC_READONLY);
|
|
|
|
/* This symbol should be passed on the command line and be variable
|
|
according to language. */
|
|
sym = symbol_new ("__gnat_vms_display_name@gnat_demangler_rtl",
|
|
absolute_section, 0, &zero_address_frag);
|
|
symbol_table_insert (sym);
|
|
symbol_get_bfdsym (sym)->flags |= BSF_DEBUGGING | BSF_DYNAMIC;
|
|
|
|
p = frag_more (4);
|
|
/* Format 3 of VMS demangler Spec. */
|
|
number_to_chars_littleendian (p, 3, 4);
|
|
|
|
p = frag_more (4);
|
|
/* Place holder for symbol table index of above symbol. */
|
|
number_to_chars_littleendian (p, -1, 4);
|
|
|
|
frag_align (3, 0, 0);
|
|
|
|
/* We probably can't restore the current segment, for there likely
|
|
isn't one yet... */
|
|
if (seg && subseg)
|
|
subseg_set (seg, subseg);
|
|
}
|
|
|
|
#endif /* TE_VMS */
|