darling-gdb/bfd/elf64-ia64-vms.c
Alan Modra 81fbe831fe PR ld/15323
bfd/
	* elf-m10300.c (mn10300_elf_check_relocs): Set non_ir_ref for
	global symbols referenced by relocs.
	* elf32-arm.c (elf32_arm_check_relocs): Likewise.
	* elf32-bfin.c (bfin_check_relocs): Likewise.
	* elf32-cr16.c (cr16_elf_check_relocs): Likewise.
	* elf32-cris.c (cris_elf_check_relocs): Likewise.
	* elf32-d10v.c (elf32_d10v_check_relocs): Likewise.
	* elf32-dlx.c (elf32_dlx_check_relocs): Likewise.
	* elf32-fr30.c (fr30_elf_check_relocs): Likewise.
	* elf32-frv.c (elf32_frv_check_relocs): Likewise.
	* elf32-hppa.c (elf32_hppa_check_relocs): Likewise.
	* elf32-i370.c (i370_elf_check_relocs): Likewise.
	* elf32-iq2000.c (iq2000_elf_check_relocs): Likewise.
	* elf32-lm32.c (lm32_elf_check_relocs): Likewise.
	* elf32-m32c.c (m32c_elf_check_relocs): Likewise.
	* elf32-m32r.c (m32r_elf_check_relocs): Likewise.
	* elf32-m68hc1x.c (elf32_m68hc11_check_relocs): Likewise.
	* elf32-m68k.c (elf_m68k_check_relocs): Likewise.
	* elf32-mcore.c (mcore_elf_check_relocs): Likewise.
	* elf32-metag.c (elf_metag_check_relocs): Likewise.
	* elf32-microblaze.c (microblaze_elf_check_relocs): Likewise.
	* elf32-moxie.c (moxie_elf_check_relocs): Likewise.
	* elf32-msp430.c (elf32_msp430_check_relocs): Likewise.
	* elf32-mt.c (mt_elf_check_relocs): Likewise.
	* elf32-nios2.c (nios2_elf32_check_relocs): Likewise.
	* elf32-openrisc.c (openrisc_elf_check_relocs): Likewise.
	* elf32-ppc.c (ppc_elf_check_relocs): Likewise.
	* elf32-rl78.c (rl78_elf_check_relocs): Likewise.
	* elf32-s390.c (elf_s390_check_relocs): Likewise.
	* elf32-score.c (s3_bfd_score_elf_check_relocs): Likewise.
	* elf32-score7.c (s7_bfd_score_elf_check_relocs): Likewise.
	* elf32-sh.c (sh_elf_check_relocs): Likewise.
	* elf32-tic6x.c (elf32_tic6x_check_relocs): Likewise.
	* elf32-tilepro.c (tilepro_elf_check_relocs): Likewise.
	* elf32-v850.c (v850_elf_check_relocs): Likewise.
	* elf32-vax.c (elf_vax_check_relocs): Likewise.
	* elf32-xstormy16.c (xstormy16_elf_check_relocs): Likewise.
	* elf32-xtensa.c (elf_xtensa_check_relocs): Likewise.
	* elf64-aarch64.c (elf64_aarch64_check_relocs): Likewise.
	* elf64-alpha.c (elf64_alpha_check_relocs): Likewise.
	* elf64-hppa.c (elf64_hppa_check_relocs): Likewise.
	* elf64-ia64-vms.c (elf64_ia64_check_relocs): Likewise.
	* elf64-mmix.c (mmix_elf_check_relocs): Likewise.
	* elf64-ppc.c (ppc64_elf_check_relocs): Likewise.
	* elf64-s390.c (elf_s390_check_relocs): Likewise.
	* elf64-sh64.c (sh_elf64_check_relocs): Likewise.
	* elfnn-ia64.c (elfNN_ia64_check_relocs): Likewise.
	* elfxx-sparc.c (_bfd_sparc_elf_check_relocs): Likewise.
	* elfxx-tilegx.c (tilegx_elf_check_relocs): Likewise.
	* elfxx-mips.c (_bfd_mips_elf_check_relocs): Likewise.  Don't
	test indirect/warning links for NULL.
ld/testsuite/
	* ld-plugin/lto.exp (pr15323a.c): Compile without -flto rather
	than using -r to effectively strip out lto info.
2013-03-30 10:14:15 +00:00

5611 lines
155 KiB
C

/* IA-64 support for OpenVMS
Copyright 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007,
2008, 2009, 2010, 2012 Free Software Foundation, Inc.
This file is part of BFD, the Binary File Descriptor library.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 51 Franklin Street - Fifth Floor, Boston,
MA 02110-1301, USA. */
#include "sysdep.h"
#include "bfd.h"
#include "libbfd.h"
#include "elf-bfd.h"
#include "opcode/ia64.h"
#include "elf/ia64.h"
#include "objalloc.h"
#include "hashtab.h"
#include "elfxx-ia64.h"
#include "vms.h"
#include "bfdver.h"
/* THE RULES for all the stuff the linker creates --
GOT Entries created in response to LTOFF or LTOFF_FPTR
relocations. Dynamic relocs created for dynamic
symbols in an application; REL relocs for locals
in a shared library.
FPTR The canonical function descriptor. Created for local
symbols in applications. Descriptors for dynamic symbols
and local symbols in shared libraries are created by
ld.so. Thus there are no dynamic relocs against these
objects. The FPTR relocs for such _are_ passed through
to the dynamic relocation tables.
FULL_PLT Created for a PCREL21B relocation against a dynamic symbol.
Requires the creation of a PLTOFF entry. This does not
require any dynamic relocations.
PLTOFF Created by PLTOFF relocations. For local symbols, this
is an alternate function descriptor, and in shared libraries
requires two REL relocations. Note that this cannot be
transformed into an FPTR relocation, since it must be in
range of the GP. For dynamic symbols, this is a function
descriptor. */
typedef struct bfd_hash_entry *(*new_hash_entry_func)
(struct bfd_hash_entry *, struct bfd_hash_table *, const char *);
/* In dynamically (linker-) created sections, we generally need to keep track
of the place a symbol or expression got allocated to. This is done via hash
tables that store entries of the following type. */
struct elf64_ia64_dyn_sym_info
{
/* The addend for which this entry is relevant. */
bfd_vma addend;
bfd_vma got_offset;
bfd_vma fptr_offset;
bfd_vma pltoff_offset;
bfd_vma plt_offset;
bfd_vma plt2_offset;
/* The symbol table entry, if any, that this was derived from. */
struct elf_link_hash_entry *h;
/* Used to count non-got, non-plt relocations for delayed sizing
of relocation sections. */
struct elf64_ia64_dyn_reloc_entry
{
struct elf64_ia64_dyn_reloc_entry *next;
asection *srel;
int type;
int count;
} *reloc_entries;
/* TRUE when the section contents have been updated. */
unsigned got_done : 1;
unsigned fptr_done : 1;
unsigned pltoff_done : 1;
/* TRUE for the different kinds of linker data we want created. */
unsigned want_got : 1;
unsigned want_gotx : 1;
unsigned want_fptr : 1;
unsigned want_ltoff_fptr : 1;
unsigned want_plt : 1; /* A MIN_PLT entry. */
unsigned want_plt2 : 1; /* A FULL_PLT. */
unsigned want_pltoff : 1;
};
struct elf64_ia64_local_hash_entry
{
int id;
unsigned int r_sym;
/* The number of elements in elf64_ia64_dyn_sym_info array. */
unsigned int count;
/* The number of sorted elements in elf64_ia64_dyn_sym_info array. */
unsigned int sorted_count;
/* The size of elf64_ia64_dyn_sym_info array. */
unsigned int size;
/* The array of elf64_ia64_dyn_sym_info. */
struct elf64_ia64_dyn_sym_info *info;
/* TRUE if this hash entry's addends was translated for
SHF_MERGE optimization. */
unsigned sec_merge_done : 1;
};
struct elf64_ia64_link_hash_entry
{
struct elf_link_hash_entry root;
/* Set if this symbol is defined in a shared library.
We can't use root.u.def.section->owner as the symbol is an absolute
symbol. */
bfd *shl;
/* The number of elements in elf64_ia64_dyn_sym_info array. */
unsigned int count;
/* The number of sorted elements in elf64_ia64_dyn_sym_info array. */
unsigned int sorted_count;
/* The size of elf64_ia64_dyn_sym_info array. */
unsigned int size;
/* The array of elf64_ia64_dyn_sym_info. */
struct elf64_ia64_dyn_sym_info *info;
};
struct elf64_ia64_link_hash_table
{
/* The main hash table. */
struct elf_link_hash_table root;
asection *fptr_sec; /* Function descriptor table (or NULL). */
asection *rel_fptr_sec; /* Dynamic relocation section for same. */
asection *pltoff_sec; /* Private descriptors for plt (or NULL). */
asection *fixups_sec; /* Fixups section. */
asection *transfer_sec; /* Transfer vector section. */
asection *note_sec; /* .note section. */
/* There are maybe R_IA64_GPREL22 relocations, including those
optimized from R_IA64_LTOFF22X, against non-SHF_IA_64_SHORT
sections. We need to record those sections so that we can choose
a proper GP to cover all R_IA64_GPREL22 relocations. */
asection *max_short_sec; /* Maximum short output section. */
bfd_vma max_short_offset; /* Maximum short offset. */
asection *min_short_sec; /* Minimum short output section. */
bfd_vma min_short_offset; /* Minimum short offset. */
htab_t loc_hash_table;
void *loc_hash_memory;
};
struct elf64_ia64_allocate_data
{
struct bfd_link_info *info;
bfd_size_type ofs;
};
#define elf64_ia64_hash_table(p) \
(elf_hash_table_id ((struct elf_link_hash_table *) ((p)->hash)) \
== IA64_ELF_DATA ? ((struct elf64_ia64_link_hash_table *) ((p)->hash)) : NULL)
struct elf64_ia64_vms_obj_tdata
{
struct elf_obj_tdata root;
/* Ident for shared library. */
bfd_uint64_t ident;
/* Used only during link: offset in the .fixups section for this bfd. */
bfd_vma fixups_off;
/* Max number of shared libraries. */
unsigned int needed_count;
};
#define elf_ia64_vms_tdata(abfd) \
((struct elf64_ia64_vms_obj_tdata *)((abfd)->tdata.any))
#define elf_ia64_vms_ident(abfd) (elf_ia64_vms_tdata(abfd)->ident)
struct elf64_vms_transfer
{
unsigned char size[4];
unsigned char spare[4];
unsigned char tfradr1[8];
unsigned char tfradr2[8];
unsigned char tfradr3[8];
unsigned char tfradr4[8];
unsigned char tfradr5[8];
/* Local function descriptor for tfr3. */
unsigned char tfr3_func[8];
unsigned char tfr3_gp[8];
};
typedef struct
{
Elf64_External_Ehdr ehdr;
unsigned char vms_needed_count[8];
} Elf64_External_VMS_Ehdr;
static struct elf64_ia64_dyn_sym_info * get_dyn_sym_info
(struct elf64_ia64_link_hash_table *,
struct elf_link_hash_entry *,
bfd *, const Elf_Internal_Rela *, bfd_boolean);
static bfd_boolean elf64_ia64_dynamic_symbol_p
(struct elf_link_hash_entry *);
static bfd_boolean elf64_ia64_choose_gp
(bfd *, struct bfd_link_info *, bfd_boolean);
static void elf64_ia64_dyn_sym_traverse
(struct elf64_ia64_link_hash_table *,
bfd_boolean (*) (struct elf64_ia64_dyn_sym_info *, void *),
void *);
static bfd_boolean allocate_global_data_got
(struct elf64_ia64_dyn_sym_info *, void *);
static bfd_boolean allocate_global_fptr_got
(struct elf64_ia64_dyn_sym_info *, void *);
static bfd_boolean allocate_local_got
(struct elf64_ia64_dyn_sym_info *, void *);
static bfd_boolean allocate_dynrel_entries
(struct elf64_ia64_dyn_sym_info *, void *);
static asection *get_pltoff
(bfd *, struct elf64_ia64_link_hash_table *);
static asection *get_got
(bfd *, struct elf64_ia64_link_hash_table *);
/* Given a ELF reloc, return the matching HOWTO structure. */
static void
elf64_ia64_info_to_howto (bfd *abfd ATTRIBUTE_UNUSED,
arelent *bfd_reloc,
Elf_Internal_Rela *elf_reloc)
{
bfd_reloc->howto
= ia64_elf_lookup_howto ((unsigned int) ELF64_R_TYPE (elf_reloc->r_info));
}
#define PLT_FULL_ENTRY_SIZE (2 * 16)
static const bfd_byte plt_full_entry[PLT_FULL_ENTRY_SIZE] =
{
0x0b, 0x78, 0x00, 0x02, 0x00, 0x24, /* [MMI] addl r15=0,r1;; */
0x00, 0x41, 0x3c, 0x70, 0x29, 0xc0, /* ld8.acq r16=[r15],8*/
0x01, 0x08, 0x00, 0x84, /* mov r14=r1;; */
0x11, 0x08, 0x00, 0x1e, 0x18, 0x10, /* [MIB] ld8 r1=[r15] */
0x60, 0x80, 0x04, 0x80, 0x03, 0x00, /* mov b6=r16 */
0x60, 0x00, 0x80, 0x00 /* br.few b6;; */
};
static const bfd_byte oor_brl[16] =
{
0x05, 0x00, 0x00, 0x00, 0x01, 0x00, /* [MLX] nop.m 0 */
0x00, 0x00, 0x00, 0x00, 0x00, 0x00, /* brl.sptk.few tgt;;*/
0x00, 0x00, 0x00, 0xc0
};
/* These functions do relaxation for IA-64 ELF. */
/* Rename some of the generic section flags to better document how they
are used here. */
#define skip_relax_pass_0 sec_flg0
#define skip_relax_pass_1 sec_flg1
static void
elf64_ia64_update_short_info (asection *sec, bfd_vma offset,
struct elf64_ia64_link_hash_table *ia64_info)
{
/* Skip ABS and SHF_IA_64_SHORT sections. */
if (sec == bfd_abs_section_ptr
|| (sec->flags & SEC_SMALL_DATA) != 0)
return;
if (!ia64_info->min_short_sec)
{
ia64_info->max_short_sec = sec;
ia64_info->max_short_offset = offset;
ia64_info->min_short_sec = sec;
ia64_info->min_short_offset = offset;
}
else if (sec == ia64_info->max_short_sec
&& offset > ia64_info->max_short_offset)
ia64_info->max_short_offset = offset;
else if (sec == ia64_info->min_short_sec
&& offset < ia64_info->min_short_offset)
ia64_info->min_short_offset = offset;
else if (sec->output_section->vma
> ia64_info->max_short_sec->vma)
{
ia64_info->max_short_sec = sec;
ia64_info->max_short_offset = offset;
}
else if (sec->output_section->vma
< ia64_info->min_short_sec->vma)
{
ia64_info->min_short_sec = sec;
ia64_info->min_short_offset = offset;
}
}
/* Use a two passes algorithm. In the first pass, branches are relaxed
(which may increase the size of the section). In the second pass,
the other relaxations are done.
*/
static bfd_boolean
elf64_ia64_relax_section (bfd *abfd, asection *sec,
struct bfd_link_info *link_info,
bfd_boolean *again)
{
struct one_fixup
{
struct one_fixup *next;
asection *tsec;
bfd_vma toff;
bfd_vma trampoff;
};
Elf_Internal_Shdr *symtab_hdr;
Elf_Internal_Rela *internal_relocs;
Elf_Internal_Rela *irel, *irelend;
bfd_byte *contents;
Elf_Internal_Sym *isymbuf = NULL;
struct elf64_ia64_link_hash_table *ia64_info;
struct one_fixup *fixups = NULL;
bfd_boolean changed_contents = FALSE;
bfd_boolean changed_relocs = FALSE;
bfd_boolean skip_relax_pass_0 = TRUE;
bfd_boolean skip_relax_pass_1 = TRUE;
bfd_vma gp = 0;
/* Assume we're not going to change any sizes, and we'll only need
one pass. */
*again = FALSE;
if (link_info->relocatable)
(*link_info->callbacks->einfo)
(_("%P%F: --relax and -r may not be used together\n"));
/* Don't even try to relax for non-ELF outputs. */
if (!is_elf_hash_table (link_info->hash))
return FALSE;
/* Nothing to do if there are no relocations or there is no need for
the current pass. */
if ((sec->flags & SEC_RELOC) == 0
|| sec->reloc_count == 0
|| (link_info->relax_pass == 0 && sec->skip_relax_pass_0)
|| (link_info->relax_pass == 1 && sec->skip_relax_pass_1))
return TRUE;
ia64_info = elf64_ia64_hash_table (link_info);
if (ia64_info == NULL)
return FALSE;
symtab_hdr = &elf_tdata (abfd)->symtab_hdr;
/* Load the relocations for this section. */
internal_relocs = (_bfd_elf_link_read_relocs
(abfd, sec, NULL, (Elf_Internal_Rela *) NULL,
link_info->keep_memory));
if (internal_relocs == NULL)
return FALSE;
irelend = internal_relocs + sec->reloc_count;
/* Get the section contents. */
if (elf_section_data (sec)->this_hdr.contents != NULL)
contents = elf_section_data (sec)->this_hdr.contents;
else
{
if (!bfd_malloc_and_get_section (abfd, sec, &contents))
goto error_return;
}
for (irel = internal_relocs; irel < irelend; irel++)
{
unsigned long r_type = ELF64_R_TYPE (irel->r_info);
bfd_vma symaddr, reladdr, trampoff, toff, roff;
asection *tsec;
struct one_fixup *f;
bfd_size_type amt;
bfd_boolean is_branch;
struct elf64_ia64_dyn_sym_info *dyn_i;
switch (r_type)
{
case R_IA64_PCREL21B:
case R_IA64_PCREL21BI:
case R_IA64_PCREL21M:
case R_IA64_PCREL21F:
/* In pass 1, all br relaxations are done. We can skip it. */
if (link_info->relax_pass == 1)
continue;
skip_relax_pass_0 = FALSE;
is_branch = TRUE;
break;
case R_IA64_PCREL60B:
/* We can't optimize brl to br in pass 0 since br relaxations
will increase the code size. Defer it to pass 1. */
if (link_info->relax_pass == 0)
{
skip_relax_pass_1 = FALSE;
continue;
}
is_branch = TRUE;
break;
case R_IA64_GPREL22:
/* Update max_short_sec/min_short_sec. */
case R_IA64_LTOFF22X:
case R_IA64_LDXMOV:
/* We can't relax ldx/mov in pass 0 since br relaxations will
increase the code size. Defer it to pass 1. */
if (link_info->relax_pass == 0)
{
skip_relax_pass_1 = FALSE;
continue;
}
is_branch = FALSE;
break;
default:
continue;
}
/* Get the value of the symbol referred to by the reloc. */
if (ELF64_R_SYM (irel->r_info) < symtab_hdr->sh_info)
{
/* A local symbol. */
Elf_Internal_Sym *isym;
/* Read this BFD's local symbols. */
if (isymbuf == NULL)
{
isymbuf = (Elf_Internal_Sym *) symtab_hdr->contents;
if (isymbuf == NULL)
isymbuf = bfd_elf_get_elf_syms (abfd, symtab_hdr,
symtab_hdr->sh_info, 0,
NULL, NULL, NULL);
if (isymbuf == 0)
goto error_return;
}
isym = isymbuf + ELF64_R_SYM (irel->r_info);
if (isym->st_shndx == SHN_UNDEF)
continue; /* We can't do anything with undefined symbols. */
else if (isym->st_shndx == SHN_ABS)
tsec = bfd_abs_section_ptr;
else if (isym->st_shndx == SHN_COMMON)
tsec = bfd_com_section_ptr;
else if (isym->st_shndx == SHN_IA_64_ANSI_COMMON)
tsec = bfd_com_section_ptr;
else
tsec = bfd_section_from_elf_index (abfd, isym->st_shndx);
toff = isym->st_value;
dyn_i = get_dyn_sym_info (ia64_info, NULL, abfd, irel, FALSE);
}
else
{
unsigned long indx;
struct elf_link_hash_entry *h;
indx = ELF64_R_SYM (irel->r_info) - symtab_hdr->sh_info;
h = elf_sym_hashes (abfd)[indx];
BFD_ASSERT (h != NULL);
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
dyn_i = get_dyn_sym_info (ia64_info, h, abfd, irel, FALSE);
/* For branches to dynamic symbols, we're interested instead
in a branch to the PLT entry. */
if (is_branch && dyn_i && dyn_i->want_plt2)
{
/* Internal branches shouldn't be sent to the PLT.
Leave this for now and we'll give an error later. */
if (r_type != R_IA64_PCREL21B)
continue;
tsec = ia64_info->root.splt;
toff = dyn_i->plt2_offset;
BFD_ASSERT (irel->r_addend == 0);
}
/* Can't do anything else with dynamic symbols. */
else if (elf64_ia64_dynamic_symbol_p (h))
continue;
else
{
/* We can't do anything with undefined symbols. */
if (h->root.type == bfd_link_hash_undefined
|| h->root.type == bfd_link_hash_undefweak)
continue;
tsec = h->root.u.def.section;
toff = h->root.u.def.value;
}
}
toff += irel->r_addend;
symaddr = tsec->output_section->vma + tsec->output_offset + toff;
roff = irel->r_offset;
if (is_branch)
{
bfd_signed_vma offset;
reladdr = (sec->output_section->vma
+ sec->output_offset
+ roff) & (bfd_vma) -4;
/* The .plt section is aligned at 32byte and the .text section
is aligned at 64byte. The .text section is right after the
.plt section. After the first relaxation pass, linker may
increase the gap between the .plt and .text sections up
to 32byte. We assume linker will always insert 32byte
between the .plt and .text sections after the first
relaxation pass. */
if (tsec == ia64_info->root.splt)
offset = -0x1000000 + 32;
else
offset = -0x1000000;
/* If the branch is in range, no need to do anything. */
if ((bfd_signed_vma) (symaddr - reladdr) >= offset
&& (bfd_signed_vma) (symaddr - reladdr) <= 0x0FFFFF0)
{
/* If the 60-bit branch is in 21-bit range, optimize it. */
if (r_type == R_IA64_PCREL60B)
{
ia64_elf_relax_brl (contents, roff);
irel->r_info = ELF64_R_INFO (ELF64_R_SYM (irel->r_info),
R_IA64_PCREL21B);
/* If the original relocation offset points to slot
1, change it to slot 2. */
if ((irel->r_offset & 3) == 1)
irel->r_offset += 1;
}
continue;
}
else if (r_type == R_IA64_PCREL60B)
continue;
else if (ia64_elf_relax_br (contents, roff))
{
irel->r_info = ELF64_R_INFO (ELF64_R_SYM (irel->r_info),
R_IA64_PCREL60B);
/* Make the relocation offset point to slot 1. */
irel->r_offset = (irel->r_offset & ~((bfd_vma) 0x3)) + 1;
continue;
}
/* We can't put a trampoline in a .init/.fini section. Issue
an error. */
if (strcmp (sec->output_section->name, ".init") == 0
|| strcmp (sec->output_section->name, ".fini") == 0)
{
(*_bfd_error_handler)
(_("%B: Can't relax br at 0x%lx in section `%A'. Please use brl or indirect branch."),
sec->owner, sec, (unsigned long) roff);
bfd_set_error (bfd_error_bad_value);
goto error_return;
}
/* If the branch and target are in the same section, you've
got one honking big section and we can't help you unless
you are branching backwards. You'll get an error message
later. */
if (tsec == sec && toff > roff)
continue;
/* Look for an existing fixup to this address. */
for (f = fixups; f ; f = f->next)
if (f->tsec == tsec && f->toff == toff)
break;
if (f == NULL)
{
/* Two alternatives: If it's a branch to a PLT entry, we can
make a copy of the FULL_PLT entry. Otherwise, we'll have
to use a `brl' insn to get where we're going. */
size_t size;
if (tsec == ia64_info->root.splt)
size = sizeof (plt_full_entry);
else
size = sizeof (oor_brl);
/* Resize the current section to make room for the new branch. */
trampoff = (sec->size + 15) & (bfd_vma) -16;
/* If trampoline is out of range, there is nothing we
can do. */
offset = trampoff - (roff & (bfd_vma) -4);
if (offset < -0x1000000 || offset > 0x0FFFFF0)
continue;
amt = trampoff + size;
contents = (bfd_byte *) bfd_realloc (contents, amt);
if (contents == NULL)
goto error_return;
sec->size = amt;
if (tsec == ia64_info->root.splt)
{
memcpy (contents + trampoff, plt_full_entry, size);
/* Hijack the old relocation for use as the PLTOFF reloc. */
irel->r_info = ELF64_R_INFO (ELF64_R_SYM (irel->r_info),
R_IA64_PLTOFF22);
irel->r_offset = trampoff;
}
else
{
memcpy (contents + trampoff, oor_brl, size);
irel->r_info = ELF64_R_INFO (ELF64_R_SYM (irel->r_info),
R_IA64_PCREL60B);
irel->r_offset = trampoff + 2;
}
/* Record the fixup so we don't do it again this section. */
f = (struct one_fixup *)
bfd_malloc ((bfd_size_type) sizeof (*f));
f->next = fixups;
f->tsec = tsec;
f->toff = toff;
f->trampoff = trampoff;
fixups = f;
}
else
{
/* If trampoline is out of range, there is nothing we
can do. */
offset = f->trampoff - (roff & (bfd_vma) -4);
if (offset < -0x1000000 || offset > 0x0FFFFF0)
continue;
/* Nop out the reloc, since we're finalizing things here. */
irel->r_info = ELF64_R_INFO (0, R_IA64_NONE);
}
/* Fix up the existing branch to hit the trampoline. */
if (ia64_elf_install_value (contents + roff, offset, r_type)
!= bfd_reloc_ok)
goto error_return;
changed_contents = TRUE;
changed_relocs = TRUE;
}
else
{
/* Fetch the gp. */
if (gp == 0)
{
bfd *obfd = sec->output_section->owner;
gp = _bfd_get_gp_value (obfd);
if (gp == 0)
{
if (!elf64_ia64_choose_gp (obfd, link_info, FALSE))
goto error_return;
gp = _bfd_get_gp_value (obfd);
}
}
/* If the data is out of range, do nothing. */
if ((bfd_signed_vma) (symaddr - gp) >= 0x200000
||(bfd_signed_vma) (symaddr - gp) < -0x200000)
continue;
if (r_type == R_IA64_GPREL22)
elf64_ia64_update_short_info (tsec->output_section,
tsec->output_offset + toff,
ia64_info);
else if (r_type == R_IA64_LTOFF22X)
{
/* Can't deal yet correctly with ABS symbols. */
if (bfd_is_abs_section (tsec))
continue;
irel->r_info = ELF64_R_INFO (ELF64_R_SYM (irel->r_info),
R_IA64_GPREL22);
changed_relocs = TRUE;
elf64_ia64_update_short_info (tsec->output_section,
tsec->output_offset + toff,
ia64_info);
}
else
{
ia64_elf_relax_ldxmov (contents, roff);
irel->r_info = ELF64_R_INFO (0, R_IA64_NONE);
changed_contents = TRUE;
changed_relocs = TRUE;
}
}
}
/* ??? If we created fixups, this may push the code segment large
enough that the data segment moves, which will change the GP.
Reset the GP so that we re-calculate next round. We need to
do this at the _beginning_ of the next round; now will not do. */
/* Clean up and go home. */
while (fixups)
{
struct one_fixup *f = fixups;
fixups = fixups->next;
free (f);
}
if (isymbuf != NULL
&& symtab_hdr->contents != (unsigned char *) isymbuf)
{
if (! link_info->keep_memory)
free (isymbuf);
else
{
/* Cache the symbols for elf_link_input_bfd. */
symtab_hdr->contents = (unsigned char *) isymbuf;
}
}
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
{
if (!changed_contents && !link_info->keep_memory)
free (contents);
else
{
/* Cache the section contents for elf_link_input_bfd. */
elf_section_data (sec)->this_hdr.contents = contents;
}
}
if (elf_section_data (sec)->relocs != internal_relocs)
{
if (!changed_relocs)
free (internal_relocs);
else
elf_section_data (sec)->relocs = internal_relocs;
}
if (link_info->relax_pass == 0)
{
/* Pass 0 is only needed to relax br. */
sec->skip_relax_pass_0 = skip_relax_pass_0;
sec->skip_relax_pass_1 = skip_relax_pass_1;
}
*again = changed_contents || changed_relocs;
return TRUE;
error_return:
if (isymbuf != NULL && (unsigned char *) isymbuf != symtab_hdr->contents)
free (isymbuf);
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
free (contents);
if (internal_relocs != NULL
&& elf_section_data (sec)->relocs != internal_relocs)
free (internal_relocs);
return FALSE;
}
#undef skip_relax_pass_0
#undef skip_relax_pass_1
/* Return TRUE if NAME is an unwind table section name. */
static inline bfd_boolean
is_unwind_section_name (bfd *abfd ATTRIBUTE_UNUSED, const char *name)
{
return ((CONST_STRNEQ (name, ELF_STRING_ia64_unwind)
&& ! CONST_STRNEQ (name, ELF_STRING_ia64_unwind_info))
|| CONST_STRNEQ (name, ELF_STRING_ia64_unwind_once));
}
/* Convert IA-64 specific section flags to bfd internal section flags. */
/* ??? There is no bfd internal flag equivalent to the SHF_IA_64_NORECOV
flag. */
static bfd_boolean
elf64_ia64_section_flags (flagword *flags,
const Elf_Internal_Shdr *hdr)
{
if (hdr->sh_flags & SHF_IA_64_SHORT)
*flags |= SEC_SMALL_DATA;
return TRUE;
}
/* Set the correct type for an IA-64 ELF section. We do this by the
section name, which is a hack, but ought to work. */
static bfd_boolean
elf64_ia64_fake_sections (bfd *abfd, Elf_Internal_Shdr *hdr,
asection *sec)
{
const char *name;
name = bfd_get_section_name (abfd, sec);
if (is_unwind_section_name (abfd, name))
{
/* We don't have the sections numbered at this point, so sh_info
is set later, in elf64_ia64_final_write_processing. */
hdr->sh_type = SHT_IA_64_UNWIND;
hdr->sh_flags |= SHF_LINK_ORDER;
}
else if (strcmp (name, ELF_STRING_ia64_archext) == 0)
hdr->sh_type = SHT_IA_64_EXT;
if (sec->flags & SEC_SMALL_DATA)
hdr->sh_flags |= SHF_IA_64_SHORT;
return TRUE;
}
/* Hook called by the linker routine which adds symbols from an object
file. We use it to put .comm items in .sbss, and not .bss. */
static bfd_boolean
elf64_ia64_add_symbol_hook (bfd *abfd,
struct bfd_link_info *info,
Elf_Internal_Sym *sym,
const char **namep ATTRIBUTE_UNUSED,
flagword *flagsp ATTRIBUTE_UNUSED,
asection **secp,
bfd_vma *valp)
{
if (sym->st_shndx == SHN_COMMON
&& !info->relocatable
&& sym->st_size <= elf_gp_size (abfd))
{
/* Common symbols less than or equal to -G nn bytes are
automatically put into .sbss. */
asection *scomm = bfd_get_section_by_name (abfd, ".scommon");
if (scomm == NULL)
{
scomm = bfd_make_section_with_flags (abfd, ".scommon",
(SEC_ALLOC
| SEC_IS_COMMON
| SEC_LINKER_CREATED));
if (scomm == NULL)
return FALSE;
}
*secp = scomm;
*valp = sym->st_size;
}
return TRUE;
}
/* According to the Tahoe assembler spec, all labels starting with a
'.' are local. */
static bfd_boolean
elf64_ia64_is_local_label_name (bfd *abfd ATTRIBUTE_UNUSED,
const char *name)
{
return name[0] == '.';
}
/* Should we do dynamic things to this symbol? */
static bfd_boolean
elf64_ia64_dynamic_symbol_p (struct elf_link_hash_entry *h)
{
return h != NULL && h->def_dynamic;
}
static struct bfd_hash_entry*
elf64_ia64_new_elf_hash_entry (struct bfd_hash_entry *entry,
struct bfd_hash_table *table,
const char *string)
{
struct elf64_ia64_link_hash_entry *ret;
ret = (struct elf64_ia64_link_hash_entry *) entry;
/* Allocate the structure if it has not already been allocated by a
subclass. */
if (!ret)
ret = bfd_hash_allocate (table, sizeof (*ret));
if (!ret)
return 0;
/* Call the allocation method of the superclass. */
ret = ((struct elf64_ia64_link_hash_entry *)
_bfd_elf_link_hash_newfunc ((struct bfd_hash_entry *) ret,
table, string));
ret->info = NULL;
ret->count = 0;
ret->sorted_count = 0;
ret->size = 0;
return (struct bfd_hash_entry *) ret;
}
static void
elf64_ia64_hash_hide_symbol (struct bfd_link_info *info,
struct elf_link_hash_entry *xh,
bfd_boolean force_local)
{
struct elf64_ia64_link_hash_entry *h;
struct elf64_ia64_dyn_sym_info *dyn_i;
unsigned int count;
h = (struct elf64_ia64_link_hash_entry *)xh;
_bfd_elf_link_hash_hide_symbol (info, &h->root, force_local);
for (count = h->count, dyn_i = h->info;
count != 0;
count--, dyn_i++)
{
dyn_i->want_plt2 = 0;
dyn_i->want_plt = 0;
}
}
/* Compute a hash of a local hash entry. */
static hashval_t
elf64_ia64_local_htab_hash (const void *ptr)
{
struct elf64_ia64_local_hash_entry *entry
= (struct elf64_ia64_local_hash_entry *) ptr;
return ELF_LOCAL_SYMBOL_HASH (entry->id, entry->r_sym);
}
/* Compare local hash entries. */
static int
elf64_ia64_local_htab_eq (const void *ptr1, const void *ptr2)
{
struct elf64_ia64_local_hash_entry *entry1
= (struct elf64_ia64_local_hash_entry *) ptr1;
struct elf64_ia64_local_hash_entry *entry2
= (struct elf64_ia64_local_hash_entry *) ptr2;
return entry1->id == entry2->id && entry1->r_sym == entry2->r_sym;
}
/* Create the derived linker hash table. The IA-64 ELF port uses this
derived hash table to keep information specific to the IA-64 ElF
linker (without using static variables). */
static struct bfd_link_hash_table *
elf64_ia64_hash_table_create (bfd *abfd)
{
struct elf64_ia64_link_hash_table *ret;
ret = bfd_zmalloc ((bfd_size_type) sizeof (*ret));
if (!ret)
return NULL;
if (!_bfd_elf_link_hash_table_init (&ret->root, abfd,
elf64_ia64_new_elf_hash_entry,
sizeof (struct elf64_ia64_link_hash_entry),
IA64_ELF_DATA))
{
free (ret);
return NULL;
}
ret->loc_hash_table = htab_try_create (1024, elf64_ia64_local_htab_hash,
elf64_ia64_local_htab_eq, NULL);
ret->loc_hash_memory = objalloc_create ();
if (!ret->loc_hash_table || !ret->loc_hash_memory)
{
free (ret);
return NULL;
}
return &ret->root.root;
}
/* Free the global elf64_ia64_dyn_sym_info array. */
static bfd_boolean
elf64_ia64_global_dyn_info_free (void **xentry,
void * unused ATTRIBUTE_UNUSED)
{
struct elf64_ia64_link_hash_entry *entry
= (struct elf64_ia64_link_hash_entry *) xentry;
if (entry->root.root.type == bfd_link_hash_warning)
entry = (struct elf64_ia64_link_hash_entry *) entry->root.root.u.i.link;
if (entry->info)
{
free (entry->info);
entry->info = NULL;
entry->count = 0;
entry->sorted_count = 0;
entry->size = 0;
}
return TRUE;
}
/* Free the local elf64_ia64_dyn_sym_info array. */
static bfd_boolean
elf64_ia64_local_dyn_info_free (void **slot,
void * unused ATTRIBUTE_UNUSED)
{
struct elf64_ia64_local_hash_entry *entry
= (struct elf64_ia64_local_hash_entry *) *slot;
if (entry->info)
{
free (entry->info);
entry->info = NULL;
entry->count = 0;
entry->sorted_count = 0;
entry->size = 0;
}
return TRUE;
}
/* Destroy IA-64 linker hash table. */
static void
elf64_ia64_hash_table_free (struct bfd_link_hash_table *hash)
{
struct elf64_ia64_link_hash_table *ia64_info
= (struct elf64_ia64_link_hash_table *) hash;
if (ia64_info->loc_hash_table)
{
htab_traverse (ia64_info->loc_hash_table,
elf64_ia64_local_dyn_info_free, NULL);
htab_delete (ia64_info->loc_hash_table);
}
if (ia64_info->loc_hash_memory)
objalloc_free ((struct objalloc *) ia64_info->loc_hash_memory);
elf_link_hash_traverse (&ia64_info->root,
elf64_ia64_global_dyn_info_free, NULL);
_bfd_elf_link_hash_table_free (hash);
}
/* Traverse both local and global hash tables. */
struct elf64_ia64_dyn_sym_traverse_data
{
bfd_boolean (*func) (struct elf64_ia64_dyn_sym_info *, void *);
void * data;
};
static bfd_boolean
elf64_ia64_global_dyn_sym_thunk (struct bfd_hash_entry *xentry,
void * xdata)
{
struct elf64_ia64_link_hash_entry *entry
= (struct elf64_ia64_link_hash_entry *) xentry;
struct elf64_ia64_dyn_sym_traverse_data *data
= (struct elf64_ia64_dyn_sym_traverse_data *) xdata;
struct elf64_ia64_dyn_sym_info *dyn_i;
unsigned int count;
if (entry->root.root.type == bfd_link_hash_warning)
entry = (struct elf64_ia64_link_hash_entry *) entry->root.root.u.i.link;
for (count = entry->count, dyn_i = entry->info;
count != 0;
count--, dyn_i++)
if (! (*data->func) (dyn_i, data->data))
return FALSE;
return TRUE;
}
static bfd_boolean
elf64_ia64_local_dyn_sym_thunk (void **slot, void * xdata)
{
struct elf64_ia64_local_hash_entry *entry
= (struct elf64_ia64_local_hash_entry *) *slot;
struct elf64_ia64_dyn_sym_traverse_data *data
= (struct elf64_ia64_dyn_sym_traverse_data *) xdata;
struct elf64_ia64_dyn_sym_info *dyn_i;
unsigned int count;
for (count = entry->count, dyn_i = entry->info;
count != 0;
count--, dyn_i++)
if (! (*data->func) (dyn_i, data->data))
return FALSE;
return TRUE;
}
static void
elf64_ia64_dyn_sym_traverse (struct elf64_ia64_link_hash_table *ia64_info,
bfd_boolean (*func) (struct elf64_ia64_dyn_sym_info *, void *),
void * data)
{
struct elf64_ia64_dyn_sym_traverse_data xdata;
xdata.func = func;
xdata.data = data;
elf_link_hash_traverse (&ia64_info->root,
elf64_ia64_global_dyn_sym_thunk, &xdata);
htab_traverse (ia64_info->loc_hash_table,
elf64_ia64_local_dyn_sym_thunk, &xdata);
}
#define NOTE_NAME "IPF/VMS"
static bfd_boolean
create_ia64_vms_notes (bfd *abfd, struct bfd_link_info *info,
unsigned int time_hi, unsigned int time_lo)
{
#define NBR_NOTES 7
Elf_Internal_Note notes[NBR_NOTES];
char *module_name;
int module_name_len;
unsigned char cur_time[8];
Elf64_External_VMS_ORIG_DYN_Note *orig_dyn;
unsigned int orig_dyn_size;
unsigned int note_size;
int i;
unsigned char *noteptr;
unsigned char *note_contents;
struct elf64_ia64_link_hash_table *ia64_info;
ia64_info = elf64_ia64_hash_table (info);
module_name = vms_get_module_name (bfd_get_filename (abfd), TRUE);
module_name_len = strlen (module_name) + 1;
bfd_putl32 (time_lo, cur_time + 0);
bfd_putl32 (time_hi, cur_time + 4);
/* Note 0: IMGNAM. */
notes[0].type = NT_VMS_IMGNAM;
notes[0].descdata = module_name;
notes[0].descsz = module_name_len;
/* Note 1: GSTNAM. */
notes[1].type = NT_VMS_GSTNAM;
notes[1].descdata = module_name;
notes[1].descsz = module_name_len;
/* Note 2: IMGID. */
#define IMG_ID "V1.0"
notes[2].type = NT_VMS_IMGID;
notes[2].descdata = IMG_ID;
notes[2].descsz = sizeof (IMG_ID);
/* Note 3: Linktime. */
notes[3].type = NT_VMS_LINKTIME;
notes[3].descdata = (char *)cur_time;
notes[3].descsz = sizeof (cur_time);
/* Note 4: Linker id. */
notes[4].type = NT_VMS_LINKID;
notes[4].descdata = "GNU ld " BFD_VERSION_STRING;
notes[4].descsz = strlen (notes[4].descdata) + 1;
/* Note 5: Original dyn. */
orig_dyn_size = (sizeof (*orig_dyn) + sizeof (IMG_ID) - 1 + 7) & ~7;
orig_dyn = bfd_zalloc (abfd, orig_dyn_size);
if (orig_dyn == NULL)
return FALSE;
bfd_putl32 (1, orig_dyn->major_id);
bfd_putl32 (3, orig_dyn->minor_id);
memcpy (orig_dyn->manipulation_date, cur_time, sizeof (cur_time));
bfd_putl64 (VMS_LF_IMGSTA | VMS_LF_MAIN, orig_dyn->link_flags);
bfd_putl32 (EF_IA_64_ABI64, orig_dyn->elf_flags);
memcpy (orig_dyn->imgid, IMG_ID, sizeof (IMG_ID));
notes[5].type = NT_VMS_ORIG_DYN;
notes[5].descdata = (char *)orig_dyn;
notes[5].descsz = orig_dyn_size;
/* Note 3: Patchtime. */
notes[6].type = NT_VMS_PATCHTIME;
notes[6].descdata = (char *)cur_time;
notes[6].descsz = sizeof (cur_time);
/* Compute notes size. */
note_size = 0;
for (i = 0; i < NBR_NOTES; i++)
note_size += sizeof (Elf64_External_VMS_Note) - 1
+ ((sizeof (NOTE_NAME) - 1 + 7) & ~7)
+ ((notes[i].descsz + 7) & ~7);
/* Malloc a temporary buffer large enough for most notes */
note_contents = (unsigned char *) bfd_zalloc (abfd, note_size);
if (note_contents == NULL)
return FALSE;
noteptr = note_contents;
/* Fill notes. */
for (i = 0; i < NBR_NOTES; i++)
{
Elf64_External_VMS_Note *enote = (Elf64_External_VMS_Note *) noteptr;
bfd_putl64 (sizeof (NOTE_NAME) - 1, enote->namesz);
bfd_putl64 (notes[i].descsz, enote->descsz);
bfd_putl64 (notes[i].type, enote->type);
noteptr = (unsigned char *)enote->name;
memcpy (noteptr, NOTE_NAME, sizeof (NOTE_NAME) - 1);
noteptr += (sizeof (NOTE_NAME) - 1 + 7) & ~7;
memcpy (noteptr, notes[i].descdata, notes[i].descsz);
noteptr += (notes[i].descsz + 7) & ~7;
}
ia64_info->note_sec->contents = note_contents;
ia64_info->note_sec->size = note_size;
free (module_name);
return TRUE;
}
static bfd_boolean
elf64_ia64_create_dynamic_sections (bfd *abfd,
struct bfd_link_info *info)
{
struct elf64_ia64_link_hash_table *ia64_info;
asection *s;
flagword flags;
const struct elf_backend_data *bed;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
if (elf_hash_table (info)->dynamic_sections_created)
return TRUE;
abfd = elf_hash_table (info)->dynobj;
bed = get_elf_backend_data (abfd);
flags = bed->dynamic_sec_flags;
s = bfd_make_section_anyway_with_flags (abfd, ".dynamic",
flags | SEC_READONLY);
if (s == NULL
|| ! bfd_set_section_alignment (abfd, s, bed->s->log_file_align))
return FALSE;
s = bfd_make_section_anyway_with_flags (abfd, ".plt", flags | SEC_READONLY);
if (s == NULL
|| ! bfd_set_section_alignment (abfd, s, bed->plt_alignment))
return FALSE;
ia64_info->root.splt = s;
if (!get_got (abfd, ia64_info))
return FALSE;
if (!get_pltoff (abfd, ia64_info))
return FALSE;
s = bfd_make_section_anyway_with_flags (abfd, ".vmsdynstr",
(SEC_ALLOC
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED));
if (s == NULL
|| !bfd_set_section_alignment (abfd, s, 0))
return FALSE;
/* Create a fixup section. */
s = bfd_make_section_anyway_with_flags (abfd, ".fixups",
(SEC_ALLOC
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED));
if (s == NULL
|| !bfd_set_section_alignment (abfd, s, 3))
return FALSE;
ia64_info->fixups_sec = s;
/* Create the transfer fixup section. */
s = bfd_make_section_anyway_with_flags (abfd, ".transfer",
(SEC_ALLOC
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED));
if (s == NULL
|| !bfd_set_section_alignment (abfd, s, 3))
return FALSE;
s->size = sizeof (struct elf64_vms_transfer);
ia64_info->transfer_sec = s;
/* Create note section. */
s = bfd_make_section_anyway_with_flags (abfd, ".vms.note",
(SEC_LINKER_CREATED
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_READONLY));
if (s == NULL
|| !bfd_set_section_alignment (abfd, s, 3))
return FALSE;
ia64_info->note_sec = s;
elf_hash_table (info)->dynamic_sections_created = TRUE;
return TRUE;
}
/* Find and/or create a hash entry for local symbol. */
static struct elf64_ia64_local_hash_entry *
get_local_sym_hash (struct elf64_ia64_link_hash_table *ia64_info,
bfd *abfd, const Elf_Internal_Rela *rel,
bfd_boolean create)
{
struct elf64_ia64_local_hash_entry e, *ret;
asection *sec = abfd->sections;
hashval_t h = ELF_LOCAL_SYMBOL_HASH (sec->id,
ELF64_R_SYM (rel->r_info));
void **slot;
e.id = sec->id;
e.r_sym = ELF64_R_SYM (rel->r_info);
slot = htab_find_slot_with_hash (ia64_info->loc_hash_table, &e, h,
create ? INSERT : NO_INSERT);
if (!slot)
return NULL;
if (*slot)
return (struct elf64_ia64_local_hash_entry *) *slot;
ret = (struct elf64_ia64_local_hash_entry *)
objalloc_alloc ((struct objalloc *) ia64_info->loc_hash_memory,
sizeof (struct elf64_ia64_local_hash_entry));
if (ret)
{
memset (ret, 0, sizeof (*ret));
ret->id = sec->id;
ret->r_sym = ELF64_R_SYM (rel->r_info);
*slot = ret;
}
return ret;
}
/* Used to sort elf64_ia64_dyn_sym_info array. */
static int
addend_compare (const void *xp, const void *yp)
{
const struct elf64_ia64_dyn_sym_info *x
= (const struct elf64_ia64_dyn_sym_info *) xp;
const struct elf64_ia64_dyn_sym_info *y
= (const struct elf64_ia64_dyn_sym_info *) yp;
return x->addend < y->addend ? -1 : x->addend > y->addend ? 1 : 0;
}
/* Sort elf64_ia64_dyn_sym_info array and remove duplicates. */
static unsigned int
sort_dyn_sym_info (struct elf64_ia64_dyn_sym_info *info,
unsigned int count)
{
bfd_vma curr, prev, got_offset;
unsigned int i, kept, dupes, diff, dest, src, len;
qsort (info, count, sizeof (*info), addend_compare);
/* Find the first duplicate. */
prev = info [0].addend;
got_offset = info [0].got_offset;
for (i = 1; i < count; i++)
{
curr = info [i].addend;
if (curr == prev)
{
/* For duplicates, make sure that GOT_OFFSET is valid. */
if (got_offset == (bfd_vma) -1)
got_offset = info [i].got_offset;
break;
}
got_offset = info [i].got_offset;
prev = curr;
}
/* We may move a block of elements to here. */
dest = i++;
/* Remove duplicates. */
if (i < count)
{
while (i < count)
{
/* For duplicates, make sure that the kept one has a valid
got_offset. */
kept = dest - 1;
if (got_offset != (bfd_vma) -1)
info [kept].got_offset = got_offset;
curr = info [i].addend;
got_offset = info [i].got_offset;
/* Move a block of elements whose first one is different from
the previous. */
if (curr == prev)
{
for (src = i + 1; src < count; src++)
{
if (info [src].addend != curr)
break;
/* For duplicates, make sure that GOT_OFFSET is
valid. */
if (got_offset == (bfd_vma) -1)
got_offset = info [src].got_offset;
}
/* Make sure that the kept one has a valid got_offset. */
if (got_offset != (bfd_vma) -1)
info [kept].got_offset = got_offset;
}
else
src = i;
if (src >= count)
break;
/* Find the next duplicate. SRC will be kept. */
prev = info [src].addend;
got_offset = info [src].got_offset;
for (dupes = src + 1; dupes < count; dupes ++)
{
curr = info [dupes].addend;
if (curr == prev)
{
/* Make sure that got_offset is valid. */
if (got_offset == (bfd_vma) -1)
got_offset = info [dupes].got_offset;
/* For duplicates, make sure that the kept one has
a valid got_offset. */
if (got_offset != (bfd_vma) -1)
info [dupes - 1].got_offset = got_offset;
break;
}
got_offset = info [dupes].got_offset;
prev = curr;
}
/* How much to move. */
len = dupes - src;
i = dupes + 1;
if (len == 1 && dupes < count)
{
/* If we only move 1 element, we combine it with the next
one. There must be at least a duplicate. Find the
next different one. */
for (diff = dupes + 1, src++; diff < count; diff++, src++)
{
if (info [diff].addend != curr)
break;
/* Make sure that got_offset is valid. */
if (got_offset == (bfd_vma) -1)
got_offset = info [diff].got_offset;
}
/* Makre sure that the last duplicated one has an valid
offset. */
BFD_ASSERT (curr == prev);
if (got_offset != (bfd_vma) -1)
info [diff - 1].got_offset = got_offset;
if (diff < count)
{
/* Find the next duplicate. Track the current valid
offset. */
prev = info [diff].addend;
got_offset = info [diff].got_offset;
for (dupes = diff + 1; dupes < count; dupes ++)
{
curr = info [dupes].addend;
if (curr == prev)
{
/* For duplicates, make sure that GOT_OFFSET
is valid. */
if (got_offset == (bfd_vma) -1)
got_offset = info [dupes].got_offset;
break;
}
got_offset = info [dupes].got_offset;
prev = curr;
diff++;
}
len = diff - src + 1;
i = diff + 1;
}
}
memmove (&info [dest], &info [src], len * sizeof (*info));
dest += len;
}
count = dest;
}
else
{
/* When we get here, either there is no duplicate at all or
the only duplicate is the last element. */
if (dest < count)
{
/* If the last element is a duplicate, make sure that the
kept one has a valid got_offset. We also update count. */
if (got_offset != (bfd_vma) -1)
info [dest - 1].got_offset = got_offset;
count = dest;
}
}
return count;
}
/* Find and/or create a descriptor for dynamic symbol info. This will
vary based on global or local symbol, and the addend to the reloc.
We don't sort when inserting. Also, we sort and eliminate
duplicates if there is an unsorted section. Typically, this will
only happen once, because we do all insertions before lookups. We
then use bsearch to do a lookup. This also allows lookups to be
fast. So we have fast insertion (O(log N) due to duplicate check),
fast lookup (O(log N)) and one sort (O(N log N) expected time).
Previously, all lookups were O(N) because of the use of the linked
list and also all insertions were O(N) because of the check for
duplicates. There are some complications here because the array
size grows occasionally, which may add an O(N) factor, but this
should be rare. Also, we free the excess array allocation, which
requires a copy which is O(N), but this only happens once. */
static struct elf64_ia64_dyn_sym_info *
get_dyn_sym_info (struct elf64_ia64_link_hash_table *ia64_info,
struct elf_link_hash_entry *h, bfd *abfd,
const Elf_Internal_Rela *rel, bfd_boolean create)
{
struct elf64_ia64_dyn_sym_info **info_p, *info, *dyn_i, key;
unsigned int *count_p, *sorted_count_p, *size_p;
unsigned int count, sorted_count, size;
bfd_vma addend = rel ? rel->r_addend : 0;
bfd_size_type amt;
if (h)
{
struct elf64_ia64_link_hash_entry *global_h;
global_h = (struct elf64_ia64_link_hash_entry *) h;
info_p = &global_h->info;
count_p = &global_h->count;
sorted_count_p = &global_h->sorted_count;
size_p = &global_h->size;
}
else
{
struct elf64_ia64_local_hash_entry *loc_h;
loc_h = get_local_sym_hash (ia64_info, abfd, rel, create);
if (!loc_h)
{
BFD_ASSERT (!create);
return NULL;
}
info_p = &loc_h->info;
count_p = &loc_h->count;
sorted_count_p = &loc_h->sorted_count;
size_p = &loc_h->size;
}
count = *count_p;
sorted_count = *sorted_count_p;
size = *size_p;
info = *info_p;
if (create)
{
/* When we create the array, we don't check for duplicates,
except in the previously sorted section if one exists, and
against the last inserted entry. This allows insertions to
be fast. */
if (info)
{
if (sorted_count)
{
/* Try bsearch first on the sorted section. */
key.addend = addend;
dyn_i = bsearch (&key, info, sorted_count,
sizeof (*info), addend_compare);
if (dyn_i)
{
return dyn_i;
}
}
/* Do a quick check for the last inserted entry. */
dyn_i = info + count - 1;
if (dyn_i->addend == addend)
{
return dyn_i;
}
}
if (size == 0)
{
/* It is the very first element. We create the array of size
1. */
size = 1;
amt = size * sizeof (*info);
info = bfd_malloc (amt);
}
else if (size <= count)
{
/* We double the array size every time when we reach the
size limit. */
size += size;
amt = size * sizeof (*info);
info = bfd_realloc (info, amt);
}
else
goto has_space;
if (info == NULL)
return NULL;
*size_p = size;
*info_p = info;
has_space:
/* Append the new one to the array. */
dyn_i = info + count;
memset (dyn_i, 0, sizeof (*dyn_i));
dyn_i->got_offset = (bfd_vma) -1;
dyn_i->addend = addend;
/* We increment count only since the new ones are unsorted and
may have duplicate. */
(*count_p)++;
}
else
{
/* It is a lookup without insertion. Sort array if part of the
array isn't sorted. */
if (count != sorted_count)
{
count = sort_dyn_sym_info (info, count);
*count_p = count;
*sorted_count_p = count;
}
/* Free unused memory. */
if (size != count)
{
amt = count * sizeof (*info);
info = bfd_malloc (amt);
if (info != NULL)
{
memcpy (info, *info_p, amt);
free (*info_p);
*size_p = count;
*info_p = info;
}
}
key.addend = addend;
dyn_i = bsearch (&key, info, count,
sizeof (*info), addend_compare);
}
return dyn_i;
}
static asection *
get_got (bfd *abfd, struct elf64_ia64_link_hash_table *ia64_info)
{
asection *got;
bfd *dynobj;
got = ia64_info->root.sgot;
if (!got)
{
flagword flags;
dynobj = ia64_info->root.dynobj;
if (!dynobj)
ia64_info->root.dynobj = dynobj = abfd;
/* The .got section is always aligned at 8 bytes. */
flags = get_elf_backend_data (dynobj)->dynamic_sec_flags;
got = bfd_make_section_anyway_with_flags (dynobj, ".got",
flags | SEC_SMALL_DATA);
if (got == NULL
|| !bfd_set_section_alignment (dynobj, got, 3))
return NULL;
ia64_info->root.sgot = got;
}
return got;
}
/* Create function descriptor section (.opd). This section is called .opd
because it contains "official procedure descriptors". The "official"
refers to the fact that these descriptors are used when taking the address
of a procedure, thus ensuring a unique address for each procedure. */
static asection *
get_fptr (bfd *abfd, struct bfd_link_info *info,
struct elf64_ia64_link_hash_table *ia64_info)
{
asection *fptr;
bfd *dynobj;
fptr = ia64_info->fptr_sec;
if (!fptr)
{
dynobj = ia64_info->root.dynobj;
if (!dynobj)
ia64_info->root.dynobj = dynobj = abfd;
fptr = bfd_make_section_anyway_with_flags (dynobj, ".opd",
(SEC_ALLOC
| SEC_LOAD
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| (info->pie ? 0
: SEC_READONLY)
| SEC_LINKER_CREATED));
if (!fptr
|| !bfd_set_section_alignment (dynobj, fptr, 4))
{
BFD_ASSERT (0);
return NULL;
}
ia64_info->fptr_sec = fptr;
if (info->pie)
{
asection *fptr_rel;
fptr_rel = bfd_make_section_anyway_with_flags (dynobj, ".rela.opd",
(SEC_ALLOC | SEC_LOAD
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED
| SEC_READONLY));
if (fptr_rel == NULL
|| !bfd_set_section_alignment (dynobj, fptr_rel, 3))
{
BFD_ASSERT (0);
return NULL;
}
ia64_info->rel_fptr_sec = fptr_rel;
}
}
return fptr;
}
static asection *
get_pltoff (bfd *abfd, struct elf64_ia64_link_hash_table *ia64_info)
{
asection *pltoff;
bfd *dynobj;
pltoff = ia64_info->pltoff_sec;
if (!pltoff)
{
dynobj = ia64_info->root.dynobj;
if (!dynobj)
ia64_info->root.dynobj = dynobj = abfd;
pltoff = bfd_make_section_anyway_with_flags (dynobj,
ELF_STRING_ia64_pltoff,
(SEC_ALLOC
| SEC_LOAD
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_SMALL_DATA
| SEC_LINKER_CREATED));
if (!pltoff
|| !bfd_set_section_alignment (dynobj, pltoff, 4))
{
BFD_ASSERT (0);
return NULL;
}
ia64_info->pltoff_sec = pltoff;
}
return pltoff;
}
static asection *
get_reloc_section (bfd *abfd,
struct elf64_ia64_link_hash_table *ia64_info,
asection *sec, bfd_boolean create)
{
const char *srel_name;
asection *srel;
bfd *dynobj;
srel_name = (bfd_elf_string_from_elf_section
(abfd, elf_elfheader(abfd)->e_shstrndx,
_bfd_elf_single_rel_hdr (sec)->sh_name));
if (srel_name == NULL)
return NULL;
BFD_ASSERT ((CONST_STRNEQ (srel_name, ".rela")
&& strcmp (bfd_get_section_name (abfd, sec),
srel_name+5) == 0)
|| (CONST_STRNEQ (srel_name, ".rel")
&& strcmp (bfd_get_section_name (abfd, sec),
srel_name+4) == 0));
dynobj = ia64_info->root.dynobj;
if (!dynobj)
ia64_info->root.dynobj = dynobj = abfd;
srel = bfd_get_linker_section (dynobj, srel_name);
if (srel == NULL && create)
{
srel = bfd_make_section_anyway_with_flags (dynobj, srel_name,
(SEC_ALLOC | SEC_LOAD
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED
| SEC_READONLY));
if (srel == NULL
|| !bfd_set_section_alignment (dynobj, srel, 3))
return NULL;
}
return srel;
}
static bfd_boolean
count_dyn_reloc (bfd *abfd, struct elf64_ia64_dyn_sym_info *dyn_i,
asection *srel, int type)
{
struct elf64_ia64_dyn_reloc_entry *rent;
for (rent = dyn_i->reloc_entries; rent; rent = rent->next)
if (rent->srel == srel && rent->type == type)
break;
if (!rent)
{
rent = ((struct elf64_ia64_dyn_reloc_entry *)
bfd_alloc (abfd, (bfd_size_type) sizeof (*rent)));
if (!rent)
return FALSE;
rent->next = dyn_i->reloc_entries;
rent->srel = srel;
rent->type = type;
rent->count = 0;
dyn_i->reloc_entries = rent;
}
rent->count++;
return TRUE;
}
static bfd_boolean
elf64_ia64_check_relocs (bfd *abfd, struct bfd_link_info *info,
asection *sec,
const Elf_Internal_Rela *relocs)
{
struct elf64_ia64_link_hash_table *ia64_info;
const Elf_Internal_Rela *relend;
Elf_Internal_Shdr *symtab_hdr;
const Elf_Internal_Rela *rel;
asection *got, *fptr, *srel, *pltoff;
enum {
NEED_GOT = 1,
NEED_GOTX = 2,
NEED_FPTR = 4,
NEED_PLTOFF = 8,
NEED_MIN_PLT = 16,
NEED_FULL_PLT = 32,
NEED_DYNREL = 64,
NEED_LTOFF_FPTR = 128
};
int need_entry;
struct elf_link_hash_entry *h;
unsigned long r_symndx;
bfd_boolean maybe_dynamic;
if (info->relocatable)
return TRUE;
symtab_hdr = &elf_tdata (abfd)->symtab_hdr;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
got = fptr = srel = pltoff = NULL;
relend = relocs + sec->reloc_count;
/* We scan relocations first to create dynamic relocation arrays. We
modified get_dyn_sym_info to allow fast insertion and support fast
lookup in the next loop. */
for (rel = relocs; rel < relend; ++rel)
{
r_symndx = ELF64_R_SYM (rel->r_info);
if (r_symndx >= symtab_hdr->sh_info)
{
long indx = r_symndx - symtab_hdr->sh_info;
h = elf_sym_hashes (abfd)[indx];
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
}
else
h = NULL;
/* We can only get preliminary data on whether a symbol is
locally or externally defined, as not all of the input files
have yet been processed. Do something with what we know, as
this may help reduce memory usage and processing time later. */
maybe_dynamic = (h && ((!info->executable
&& (!SYMBOLIC_BIND (info, h)
|| info->unresolved_syms_in_shared_libs == RM_IGNORE))
|| !h->def_regular
|| h->root.type == bfd_link_hash_defweak));
need_entry = 0;
switch (ELF64_R_TYPE (rel->r_info))
{
case R_IA64_TPREL64MSB:
case R_IA64_TPREL64LSB:
case R_IA64_LTOFF_TPREL22:
case R_IA64_DTPREL32MSB:
case R_IA64_DTPREL32LSB:
case R_IA64_DTPREL64MSB:
case R_IA64_DTPREL64LSB:
case R_IA64_LTOFF_DTPREL22:
case R_IA64_DTPMOD64MSB:
case R_IA64_DTPMOD64LSB:
case R_IA64_LTOFF_DTPMOD22:
abort ();
break;
case R_IA64_IPLTMSB:
case R_IA64_IPLTLSB:
break;
case R_IA64_LTOFF_FPTR22:
case R_IA64_LTOFF_FPTR64I:
case R_IA64_LTOFF_FPTR32MSB:
case R_IA64_LTOFF_FPTR32LSB:
case R_IA64_LTOFF_FPTR64MSB:
case R_IA64_LTOFF_FPTR64LSB:
need_entry = NEED_FPTR | NEED_GOT | NEED_LTOFF_FPTR;
break;
case R_IA64_FPTR64I:
case R_IA64_FPTR32MSB:
case R_IA64_FPTR32LSB:
case R_IA64_FPTR64MSB:
case R_IA64_FPTR64LSB:
if (info->shared || h)
need_entry = NEED_FPTR | NEED_DYNREL;
else
need_entry = NEED_FPTR;
break;
case R_IA64_LTOFF22:
case R_IA64_LTOFF64I:
need_entry = NEED_GOT;
break;
case R_IA64_LTOFF22X:
need_entry = NEED_GOTX;
break;
case R_IA64_PLTOFF22:
case R_IA64_PLTOFF64I:
case R_IA64_PLTOFF64MSB:
case R_IA64_PLTOFF64LSB:
need_entry = NEED_PLTOFF;
if (h)
{
if (maybe_dynamic)
need_entry |= NEED_MIN_PLT;
}
else
{
(*info->callbacks->warning)
(info, _("@pltoff reloc against local symbol"), 0,
abfd, 0, (bfd_vma) 0);
}
break;
case R_IA64_PCREL21B:
case R_IA64_PCREL60B:
/* Depending on where this symbol is defined, we may or may not
need a full plt entry. Only skip if we know we'll not need
the entry -- static or symbolic, and the symbol definition
has already been seen. */
if (maybe_dynamic && rel->r_addend == 0)
need_entry = NEED_FULL_PLT;
break;
case R_IA64_IMM14:
case R_IA64_IMM22:
case R_IA64_IMM64:
case R_IA64_DIR32MSB:
case R_IA64_DIR32LSB:
case R_IA64_DIR64MSB:
case R_IA64_DIR64LSB:
/* Shared objects will always need at least a REL relocation. */
if (info->shared || maybe_dynamic)
need_entry = NEED_DYNREL;
break;
case R_IA64_PCREL22:
case R_IA64_PCREL64I:
case R_IA64_PCREL32MSB:
case R_IA64_PCREL32LSB:
case R_IA64_PCREL64MSB:
case R_IA64_PCREL64LSB:
if (maybe_dynamic)
need_entry = NEED_DYNREL;
break;
}
if (!need_entry)
continue;
if ((need_entry & NEED_FPTR) != 0
&& rel->r_addend)
{
(*info->callbacks->warning)
(info, _("non-zero addend in @fptr reloc"), 0,
abfd, 0, (bfd_vma) 0);
}
if (get_dyn_sym_info (ia64_info, h, abfd, rel, TRUE) == NULL)
return FALSE;
}
/* Now, we only do lookup without insertion, which is very fast
with the modified get_dyn_sym_info. */
for (rel = relocs; rel < relend; ++rel)
{
struct elf64_ia64_dyn_sym_info *dyn_i;
int dynrel_type = R_IA64_NONE;
r_symndx = ELF64_R_SYM (rel->r_info);
if (r_symndx >= symtab_hdr->sh_info)
{
/* We're dealing with a global symbol -- find its hash entry
and mark it as being referenced. */
long indx = r_symndx - symtab_hdr->sh_info;
h = elf_sym_hashes (abfd)[indx];
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
/* PR15323, ref flags aren't set for references in the same
object. */
h->root.non_ir_ref = 1;
h->ref_regular = 1;
}
else
h = NULL;
/* We can only get preliminary data on whether a symbol is
locally or externally defined, as not all of the input files
have yet been processed. Do something with what we know, as
this may help reduce memory usage and processing time later. */
maybe_dynamic = (h && ((!info->executable
&& (!SYMBOLIC_BIND (info, h)
|| info->unresolved_syms_in_shared_libs == RM_IGNORE))
|| !h->def_regular
|| h->root.type == bfd_link_hash_defweak));
need_entry = 0;
switch (ELF64_R_TYPE (rel->r_info))
{
case R_IA64_TPREL64MSB:
case R_IA64_TPREL64LSB:
case R_IA64_LTOFF_TPREL22:
case R_IA64_DTPREL32MSB:
case R_IA64_DTPREL32LSB:
case R_IA64_DTPREL64MSB:
case R_IA64_DTPREL64LSB:
case R_IA64_LTOFF_DTPREL22:
case R_IA64_DTPMOD64MSB:
case R_IA64_DTPMOD64LSB:
case R_IA64_LTOFF_DTPMOD22:
abort ();
break;
case R_IA64_LTOFF_FPTR22:
case R_IA64_LTOFF_FPTR64I:
case R_IA64_LTOFF_FPTR32MSB:
case R_IA64_LTOFF_FPTR32LSB:
case R_IA64_LTOFF_FPTR64MSB:
case R_IA64_LTOFF_FPTR64LSB:
need_entry = NEED_FPTR | NEED_GOT | NEED_LTOFF_FPTR;
break;
case R_IA64_FPTR64I:
case R_IA64_FPTR32MSB:
case R_IA64_FPTR32LSB:
case R_IA64_FPTR64MSB:
case R_IA64_FPTR64LSB:
if (info->shared || h)
need_entry = NEED_FPTR | NEED_DYNREL;
else
need_entry = NEED_FPTR;
dynrel_type = R_IA64_FPTR64LSB;
break;
case R_IA64_LTOFF22:
case R_IA64_LTOFF64I:
need_entry = NEED_GOT;
break;
case R_IA64_LTOFF22X:
need_entry = NEED_GOTX;
break;
case R_IA64_PLTOFF22:
case R_IA64_PLTOFF64I:
case R_IA64_PLTOFF64MSB:
case R_IA64_PLTOFF64LSB:
need_entry = NEED_PLTOFF;
if (h)
{
if (maybe_dynamic)
need_entry |= NEED_MIN_PLT;
}
break;
case R_IA64_PCREL21B:
case R_IA64_PCREL60B:
/* Depending on where this symbol is defined, we may or may not
need a full plt entry. Only skip if we know we'll not need
the entry -- static or symbolic, and the symbol definition
has already been seen. */
if (maybe_dynamic && rel->r_addend == 0)
need_entry = NEED_FULL_PLT;
break;
case R_IA64_IMM14:
case R_IA64_IMM22:
case R_IA64_IMM64:
case R_IA64_DIR32MSB:
case R_IA64_DIR32LSB:
case R_IA64_DIR64MSB:
case R_IA64_DIR64LSB:
/* Shared objects will always need at least a REL relocation. */
if (info->shared || maybe_dynamic)
need_entry = NEED_DYNREL;
dynrel_type = R_IA64_DIR64LSB;
break;
case R_IA64_IPLTMSB:
case R_IA64_IPLTLSB:
break;
case R_IA64_PCREL22:
case R_IA64_PCREL64I:
case R_IA64_PCREL32MSB:
case R_IA64_PCREL32LSB:
case R_IA64_PCREL64MSB:
case R_IA64_PCREL64LSB:
if (maybe_dynamic)
need_entry = NEED_DYNREL;
dynrel_type = R_IA64_PCREL64LSB;
break;
}
if (!need_entry)
continue;
dyn_i = get_dyn_sym_info (ia64_info, h, abfd, rel, FALSE);
/* Record whether or not this is a local symbol. */
dyn_i->h = h;
/* Create what's needed. */
if (need_entry & (NEED_GOT | NEED_GOTX))
{
if (!got)
{
got = get_got (abfd, ia64_info);
if (!got)
return FALSE;
}
if (need_entry & NEED_GOT)
dyn_i->want_got = 1;
if (need_entry & NEED_GOTX)
dyn_i->want_gotx = 1;
}
if (need_entry & NEED_FPTR)
{
/* Create the .opd section. */
if (!fptr)
{
fptr = get_fptr (abfd, info, ia64_info);
if (!fptr)
return FALSE;
}
dyn_i->want_fptr = 1;
}
if (need_entry & NEED_LTOFF_FPTR)
dyn_i->want_ltoff_fptr = 1;
if (need_entry & (NEED_MIN_PLT | NEED_FULL_PLT))
{
if (!ia64_info->root.dynobj)
ia64_info->root.dynobj = abfd;
h->needs_plt = 1;
dyn_i->want_plt = 1;
}
if (need_entry & NEED_FULL_PLT)
dyn_i->want_plt2 = 1;
if (need_entry & NEED_PLTOFF)
{
/* This is needed here, in case @pltoff is used in a non-shared
link. */
if (!pltoff)
{
pltoff = get_pltoff (abfd, ia64_info);
if (!pltoff)
return FALSE;
}
dyn_i->want_pltoff = 1;
}
if ((need_entry & NEED_DYNREL) && (sec->flags & SEC_ALLOC))
{
if (!srel)
{
srel = get_reloc_section (abfd, ia64_info, sec, TRUE);
if (!srel)
return FALSE;
}
if (!count_dyn_reloc (abfd, dyn_i, srel, dynrel_type))
return FALSE;
}
}
return TRUE;
}
/* For cleanliness, and potentially faster dynamic loading, allocate
external GOT entries first. */
static bfd_boolean
allocate_global_data_got (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *)data;
if ((dyn_i->want_got || dyn_i->want_gotx)
&& ! dyn_i->want_fptr
&& elf64_ia64_dynamic_symbol_p (dyn_i->h))
{
/* GOT entry with FPTR is done by allocate_global_fptr_got. */
dyn_i->got_offset = x->ofs;
x->ofs += 8;
}
return TRUE;
}
/* Next, allocate all the GOT entries used by LTOFF_FPTR relocs. */
static bfd_boolean
allocate_global_fptr_got (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *)data;
if (dyn_i->want_got
&& dyn_i->want_fptr
&& elf64_ia64_dynamic_symbol_p (dyn_i->h))
{
dyn_i->got_offset = x->ofs;
x->ofs += 8;
}
return TRUE;
}
/* Lastly, allocate all the GOT entries for local data. */
static bfd_boolean
allocate_local_got (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *) data;
if ((dyn_i->want_got || dyn_i->want_gotx)
&& !elf64_ia64_dynamic_symbol_p (dyn_i->h))
{
dyn_i->got_offset = x->ofs;
x->ofs += 8;
}
return TRUE;
}
/* Allocate function descriptors. We can do these for every function
in a main executable that is not exported. */
static bfd_boolean
allocate_fptr (struct elf64_ia64_dyn_sym_info *dyn_i, void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *) data;
if (dyn_i->want_fptr)
{
struct elf_link_hash_entry *h = dyn_i->h;
if (h)
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
if (h == NULL || !h->def_dynamic)
{
/* A non dynamic symbol. */
dyn_i->fptr_offset = x->ofs;
x->ofs += 16;
}
else
dyn_i->want_fptr = 0;
}
return TRUE;
}
/* Allocate all the minimal PLT entries. */
static bfd_boolean
allocate_plt_entries (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data ATTRIBUTE_UNUSED)
{
if (dyn_i->want_plt)
{
struct elf_link_hash_entry *h = dyn_i->h;
if (h)
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
/* ??? Versioned symbols seem to lose NEEDS_PLT. */
if (elf64_ia64_dynamic_symbol_p (h))
{
dyn_i->want_pltoff = 1;
}
else
{
dyn_i->want_plt = 0;
dyn_i->want_plt2 = 0;
}
}
return TRUE;
}
/* Allocate all the full PLT entries. */
static bfd_boolean
allocate_plt2_entries (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *)data;
if (dyn_i->want_plt2)
{
struct elf_link_hash_entry *h = dyn_i->h;
bfd_size_type ofs = x->ofs;
dyn_i->plt2_offset = ofs;
x->ofs = ofs + PLT_FULL_ENTRY_SIZE;
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
dyn_i->h->plt.offset = ofs;
}
return TRUE;
}
/* Allocate all the PLTOFF entries requested by relocations and
plt entries. We can't share space with allocated FPTR entries,
because the latter are not necessarily addressable by the GP.
??? Relaxation might be able to determine that they are. */
static bfd_boolean
allocate_pltoff_entries (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *)data;
if (dyn_i->want_pltoff)
{
dyn_i->pltoff_offset = x->ofs;
x->ofs += 16;
}
return TRUE;
}
/* Allocate dynamic relocations for those symbols that turned out
to be dynamic. */
static bfd_boolean
allocate_dynrel_entries (struct elf64_ia64_dyn_sym_info *dyn_i,
void * data)
{
struct elf64_ia64_allocate_data *x = (struct elf64_ia64_allocate_data *)data;
struct elf64_ia64_link_hash_table *ia64_info;
struct elf64_ia64_dyn_reloc_entry *rent;
bfd_boolean dynamic_symbol, shared, resolved_zero;
struct elf64_ia64_link_hash_entry *h_ia64;
ia64_info = elf64_ia64_hash_table (x->info);
if (ia64_info == NULL)
return FALSE;
/* Note that this can't be used in relation to FPTR relocs below. */
dynamic_symbol = elf64_ia64_dynamic_symbol_p (dyn_i->h);
shared = x->info->shared;
resolved_zero = (dyn_i->h
&& ELF_ST_VISIBILITY (dyn_i->h->other)
&& dyn_i->h->root.type == bfd_link_hash_undefweak);
/* Take care of the GOT and PLT relocations. */
if ((!resolved_zero
&& (dynamic_symbol || shared)
&& (dyn_i->want_got || dyn_i->want_gotx))
|| (dyn_i->want_ltoff_fptr
&& dyn_i->h
&& dyn_i->h->def_dynamic))
{
/* VMS: FIX64. */
if (dyn_i->h != NULL && dyn_i->h->def_dynamic)
{
h_ia64 = (struct elf64_ia64_link_hash_entry *) dyn_i->h;
elf_ia64_vms_tdata (h_ia64->shl)->fixups_off +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
ia64_info->fixups_sec->size +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
}
}
if (ia64_info->rel_fptr_sec && dyn_i->want_fptr)
{
/* VMS: only image reloc. */
if (dyn_i->h == NULL || dyn_i->h->root.type != bfd_link_hash_undefweak)
ia64_info->rel_fptr_sec->size += sizeof (Elf64_External_Rela);
}
if (!resolved_zero && dyn_i->want_pltoff)
{
/* VMS: FIXFD. */
if (dyn_i->h != NULL && dyn_i->h->def_dynamic)
{
h_ia64 = (struct elf64_ia64_link_hash_entry *) dyn_i->h;
elf_ia64_vms_tdata (h_ia64->shl)->fixups_off +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
ia64_info->fixups_sec->size +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
}
}
/* Take care of the normal data relocations. */
for (rent = dyn_i->reloc_entries; rent; rent = rent->next)
{
int count = rent->count;
switch (rent->type)
{
case R_IA64_FPTR32LSB:
case R_IA64_FPTR64LSB:
/* Allocate one iff !want_fptr and not PIE, which by this point
will be true only if we're actually allocating one statically
in the main executable. Position independent executables
need a relative reloc. */
if (dyn_i->want_fptr && !x->info->pie)
continue;
break;
case R_IA64_PCREL32LSB:
case R_IA64_PCREL64LSB:
if (!dynamic_symbol)
continue;
break;
case R_IA64_DIR32LSB:
case R_IA64_DIR64LSB:
if (!dynamic_symbol && !shared)
continue;
break;
case R_IA64_IPLTLSB:
if (!dynamic_symbol && !shared)
continue;
/* Use two REL relocations for IPLT relocations
against local symbols. */
if (!dynamic_symbol)
count *= 2;
break;
case R_IA64_DTPREL32LSB:
case R_IA64_TPREL64LSB:
case R_IA64_DTPREL64LSB:
case R_IA64_DTPMOD64LSB:
break;
default:
abort ();
}
/* Add a fixup. */
if (!dynamic_symbol)
abort ();
h_ia64 = (struct elf64_ia64_link_hash_entry *) dyn_i->h;
elf_ia64_vms_tdata (h_ia64->shl)->fixups_off +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
ia64_info->fixups_sec->size +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
}
return TRUE;
}
static bfd_boolean
elf64_ia64_adjust_dynamic_symbol (struct bfd_link_info *info ATTRIBUTE_UNUSED,
struct elf_link_hash_entry *h)
{
/* ??? Undefined symbols with PLT entries should be re-defined
to be the PLT entry. */
/* If this is a weak symbol, and there is a real definition, the
processor independent code will have arranged for us to see the
real definition first, and we can just use the same value. */
if (h->u.weakdef != NULL)
{
BFD_ASSERT (h->u.weakdef->root.type == bfd_link_hash_defined
|| h->u.weakdef->root.type == bfd_link_hash_defweak);
h->root.u.def.section = h->u.weakdef->root.u.def.section;
h->root.u.def.value = h->u.weakdef->root.u.def.value;
return TRUE;
}
/* If this is a reference to a symbol defined by a dynamic object which
is not a function, we might allocate the symbol in our .dynbss section
and allocate a COPY dynamic relocation.
But IA-64 code is canonically PIC, so as a rule we can avoid this sort
of hackery. */
return TRUE;
}
static bfd_boolean
elf64_ia64_size_dynamic_sections (bfd *output_bfd ATTRIBUTE_UNUSED,
struct bfd_link_info *info)
{
struct elf64_ia64_allocate_data data;
struct elf64_ia64_link_hash_table *ia64_info;
asection *sec;
bfd *dynobj;
struct elf_link_hash_table *hash_table;
hash_table = elf_hash_table (info);
dynobj = hash_table->dynobj;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
BFD_ASSERT(dynobj != NULL);
data.info = info;
/* Allocate the GOT entries. */
if (ia64_info->root.sgot)
{
data.ofs = 0;
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_global_data_got, &data);
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_global_fptr_got, &data);
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_local_got, &data);
ia64_info->root.sgot->size = data.ofs;
}
/* Allocate the FPTR entries. */
if (ia64_info->fptr_sec)
{
data.ofs = 0;
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_fptr, &data);
ia64_info->fptr_sec->size = data.ofs;
}
/* Now that we've seen all of the input files, we can decide which
symbols need plt entries. Allocate the minimal PLT entries first.
We do this even though dynamic_sections_created may be FALSE, because
this has the side-effect of clearing want_plt and want_plt2. */
data.ofs = 0;
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_plt_entries, &data);
/* Align the pointer for the plt2 entries. */
data.ofs = (data.ofs + 31) & (bfd_vma) -32;
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_plt2_entries, &data);
if (data.ofs != 0 || ia64_info->root.dynamic_sections_created)
{
/* FIXME: we always reserve the memory for dynamic linker even if
there are no PLT entries since dynamic linker may assume the
reserved memory always exists. */
BFD_ASSERT (ia64_info->root.dynamic_sections_created);
ia64_info->root.splt->size = data.ofs;
}
/* Allocate the PLTOFF entries. */
if (ia64_info->pltoff_sec)
{
data.ofs = 0;
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_pltoff_entries, &data);
ia64_info->pltoff_sec->size = data.ofs;
}
if (ia64_info->root.dynamic_sections_created)
{
/* Allocate space for the dynamic relocations that turned out to be
required. */
elf64_ia64_dyn_sym_traverse (ia64_info, allocate_dynrel_entries, &data);
}
/* We have now determined the sizes of the various dynamic sections.
Allocate memory for them. */
for (sec = dynobj->sections; sec != NULL; sec = sec->next)
{
bfd_boolean strip;
if (!(sec->flags & SEC_LINKER_CREATED))
continue;
/* If we don't need this section, strip it from the output file.
There were several sections primarily related to dynamic
linking that must be create before the linker maps input
sections to output sections. The linker does that before
bfd_elf_size_dynamic_sections is called, and it is that
function which decides whether anything needs to go into
these sections. */
strip = (sec->size == 0);
if (sec == ia64_info->root.sgot)
strip = FALSE;
else if (sec == ia64_info->root.srelgot)
{
if (strip)
ia64_info->root.srelgot = NULL;
else
/* We use the reloc_count field as a counter if we need to
copy relocs into the output file. */
sec->reloc_count = 0;
}
else if (sec == ia64_info->fptr_sec)
{
if (strip)
ia64_info->fptr_sec = NULL;
}
else if (sec == ia64_info->rel_fptr_sec)
{
if (strip)
ia64_info->rel_fptr_sec = NULL;
else
/* We use the reloc_count field as a counter if we need to
copy relocs into the output file. */
sec->reloc_count = 0;
}
else if (sec == ia64_info->root.splt)
{
if (strip)
ia64_info->root.splt = NULL;
}
else if (sec == ia64_info->pltoff_sec)
{
if (strip)
ia64_info->pltoff_sec = NULL;
}
else if (sec == ia64_info->fixups_sec)
{
if (strip)
ia64_info->fixups_sec = NULL;
}
else if (sec == ia64_info->transfer_sec)
{
;
}
else
{
const char *name;
/* It's OK to base decisions on the section name, because none
of the dynobj section names depend upon the input files. */
name = bfd_get_section_name (dynobj, sec);
if (strcmp (name, ".got.plt") == 0)
strip = FALSE;
else if (CONST_STRNEQ (name, ".rel"))
{
if (!strip)
{
/* We use the reloc_count field as a counter if we need to
copy relocs into the output file. */
sec->reloc_count = 0;
}
}
else
continue;
}
if (strip)
sec->flags |= SEC_EXCLUDE;
else
{
/* Allocate memory for the section contents. */
sec->contents = (bfd_byte *) bfd_zalloc (dynobj, sec->size);
if (sec->contents == NULL && sec->size != 0)
return FALSE;
}
}
if (elf_hash_table (info)->dynamic_sections_created)
{
bfd *abfd;
asection *dynsec;
asection *dynstrsec;
Elf_Internal_Dyn dyn;
const struct elf_backend_data *bed;
unsigned int shl_num = 0;
bfd_vma fixups_off = 0;
bfd_vma strdyn_off;
unsigned int time_hi, time_lo;
/* The .dynamic section must exist and be empty. */
dynsec = bfd_get_linker_section (hash_table->dynobj, ".dynamic");
BFD_ASSERT (dynsec != NULL);
BFD_ASSERT (dynsec->size == 0);
dynstrsec = bfd_get_linker_section (hash_table->dynobj, ".vmsdynstr");
BFD_ASSERT (dynstrsec != NULL);
BFD_ASSERT (dynstrsec->size == 0);
dynstrsec->size = 1; /* Initial blank. */
/* Ident + link time. */
vms_get_time (&time_hi, &time_lo);
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_IDENT, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_LINKTIME,
(((bfd_uint64_t)time_hi) << 32)
+ time_lo))
return FALSE;
/* Strtab. */
strdyn_off = dynsec->size;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_STRTAB_OFFSET, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_STRSZ, 0))
return FALSE;
/* PLTGOT */
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_PLTGOT_SEG, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_PLTGOT_OFFSET, 0))
return FALSE;
/* Misc. */
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_FPMODE, 0x9800000))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_LNKFLAGS,
VMS_LF_IMGSTA | VMS_LF_MAIN))
return FALSE;
/* Add entries for shared libraries. */
for (abfd = info->input_bfds; abfd; abfd = abfd->link_next)
{
char *soname;
size_t soname_len;
bfd_size_type strindex;
bfd_byte *newcontents;
bfd_vma fixups_shl_off;
if (!(abfd->flags & DYNAMIC))
continue;
BFD_ASSERT (abfd->xvec == output_bfd->xvec);
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_NEEDED_IDENT,
elf_ia64_vms_ident (abfd)))
return FALSE;
soname = vms_get_module_name (abfd->filename, TRUE);
if (soname == NULL)
return FALSE;
strindex = dynstrsec->size;
soname_len = strlen (soname) + 1;
newcontents = (bfd_byte *) bfd_realloc (dynstrsec->contents,
strindex + soname_len);
if (newcontents == NULL)
return FALSE;
memcpy (newcontents + strindex, soname, soname_len);
dynstrsec->size += soname_len;
dynstrsec->contents = newcontents;
if (!_bfd_elf_add_dynamic_entry (info, DT_NEEDED, strindex))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_FIXUP_NEEDED,
shl_num))
return FALSE;
shl_num++;
/* The fixups_off was in fact containing the size of the fixup
section. Remap into the offset. */
fixups_shl_off = elf_ia64_vms_tdata (abfd)->fixups_off;
elf_ia64_vms_tdata (abfd)->fixups_off = fixups_off;
if (!_bfd_elf_add_dynamic_entry
(info, DT_IA_64_VMS_FIXUP_RELA_CNT,
fixups_shl_off / sizeof (Elf64_External_VMS_IMAGE_FIXUP)))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_FIXUP_RELA_OFF,
fixups_off))
return FALSE;
fixups_off += fixups_shl_off;
}
/* Unwind. */
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_UNWINDSZ, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_UNWIND_CODSEG, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_UNWIND_INFOSEG, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_UNWIND_OFFSET, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_IA_64_VMS_UNWIND_SEG, 0))
return FALSE;
if (!_bfd_elf_add_dynamic_entry (info, DT_NULL, 0xdead))
return FALSE;
/* Fix the strtab entries. */
bed = get_elf_backend_data (hash_table->dynobj);
if (dynstrsec->size > 1)
dynstrsec->contents[0] = 0;
else
dynstrsec->size = 0;
/* Note: one 'spare' (ie DT_NULL) entry is added by
bfd_elf_size_dynsym_hash_dynstr. */
dyn.d_tag = DT_IA_64_VMS_STRTAB_OFFSET;
dyn.d_un.d_val = dynsec->size /* + sizeof (Elf64_External_Dyn) */;
bed->s->swap_dyn_out (hash_table->dynobj, &dyn,
dynsec->contents + strdyn_off);
dyn.d_tag = DT_STRSZ;
dyn.d_un.d_val = dynstrsec->size;
bed->s->swap_dyn_out (hash_table->dynobj, &dyn,
dynsec->contents + strdyn_off + bed->s->sizeof_dyn);
elf_ia64_vms_tdata (output_bfd)->needed_count = shl_num;
/* Note section. */
if (!create_ia64_vms_notes (output_bfd, info, time_hi, time_lo))
return FALSE;
}
/* ??? Perhaps force __gp local. */
return TRUE;
}
static void
elf64_ia64_install_fixup (bfd *output_bfd,
struct elf64_ia64_link_hash_table *ia64_info,
struct elf_link_hash_entry *h,
unsigned int type, asection *sec, bfd_vma offset,
bfd_vma addend)
{
asection *relsec;
Elf64_External_VMS_IMAGE_FIXUP *fixup;
struct elf64_ia64_link_hash_entry *h_ia64;
bfd_vma fixoff;
Elf_Internal_Phdr *phdr;
if (h == NULL || !h->def_dynamic)
abort ();
h_ia64 = (struct elf64_ia64_link_hash_entry *) h;
fixoff = elf_ia64_vms_tdata (h_ia64->shl)->fixups_off;
elf_ia64_vms_tdata (h_ia64->shl)->fixups_off +=
sizeof (Elf64_External_VMS_IMAGE_FIXUP);
relsec = ia64_info->fixups_sec;
fixup = (Elf64_External_VMS_IMAGE_FIXUP *)(relsec->contents + fixoff);
offset += sec->output_section->vma + sec->output_offset;
/* FIXME: this is slow. We should cache the last one used, or create a
map. */
phdr = _bfd_elf_find_segment_containing_section
(output_bfd, sec->output_section);
BFD_ASSERT (phdr != NULL);
bfd_putl64 (offset - phdr->p_vaddr, fixup->fixup_offset);
bfd_putl32 (type, fixup->type);
bfd_putl32 (phdr - elf_tdata (output_bfd)->phdr, fixup->fixup_seg);
bfd_putl64 (addend, fixup->addend);
bfd_putl32 (h->root.u.def.value, fixup->symvec_index);
bfd_putl32 (2, fixup->data_type);
}
/* Store an entry for target address TARGET_ADDR in the linkage table
and return the gp-relative address of the linkage table entry. */
static bfd_vma
set_got_entry (bfd *abfd, struct bfd_link_info *info,
struct elf64_ia64_dyn_sym_info *dyn_i,
bfd_vma addend, bfd_vma value, unsigned int dyn_r_type)
{
struct elf64_ia64_link_hash_table *ia64_info;
asection *got_sec;
bfd_boolean done;
bfd_vma got_offset;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return 0;
got_sec = ia64_info->root.sgot;
switch (dyn_r_type)
{
case R_IA64_TPREL64LSB:
case R_IA64_DTPMOD64LSB:
case R_IA64_DTPREL32LSB:
case R_IA64_DTPREL64LSB:
abort ();
break;
default:
done = dyn_i->got_done;
dyn_i->got_done = TRUE;
got_offset = dyn_i->got_offset;
break;
}
BFD_ASSERT ((got_offset & 7) == 0);
if (! done)
{
/* Store the target address in the linkage table entry. */
bfd_put_64 (abfd, value, got_sec->contents + got_offset);
/* Install a dynamic relocation if needed. */
if (((info->shared
&& (!dyn_i->h
|| ELF_ST_VISIBILITY (dyn_i->h->other) == STV_DEFAULT
|| dyn_i->h->root.type != bfd_link_hash_undefweak))
|| elf64_ia64_dynamic_symbol_p (dyn_i->h))
&& (!dyn_i->want_ltoff_fptr
|| !info->pie
|| !dyn_i->h
|| dyn_i->h->root.type != bfd_link_hash_undefweak))
{
if (!dyn_i->h || !dyn_i->h->def_dynamic)
{
dyn_r_type = R_IA64_REL64LSB;
addend = value;
}
/* VMS: install a FIX32 or FIX64. */
switch (dyn_r_type)
{
case R_IA64_DIR32LSB:
case R_IA64_FPTR32LSB:
dyn_r_type = R_IA64_VMS_FIX32;
break;
case R_IA64_DIR64LSB:
case R_IA64_FPTR64LSB:
dyn_r_type = R_IA64_VMS_FIX64;
break;
default:
BFD_ASSERT (FALSE);
break;
}
elf64_ia64_install_fixup
(info->output_bfd, ia64_info, dyn_i->h,
dyn_r_type, got_sec, got_offset, addend);
}
}
/* Return the address of the linkage table entry. */
value = (got_sec->output_section->vma
+ got_sec->output_offset
+ got_offset);
return value;
}
/* Fill in a function descriptor consisting of the function's code
address and its global pointer. Return the descriptor's address. */
static bfd_vma
set_fptr_entry (bfd *abfd, struct bfd_link_info *info,
struct elf64_ia64_dyn_sym_info *dyn_i,
bfd_vma value)
{
struct elf64_ia64_link_hash_table *ia64_info;
asection *fptr_sec;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return 0;
fptr_sec = ia64_info->fptr_sec;
if (!dyn_i->fptr_done)
{
dyn_i->fptr_done = 1;
/* Fill in the function descriptor. */
bfd_put_64 (abfd, value, fptr_sec->contents + dyn_i->fptr_offset);
bfd_put_64 (abfd, _bfd_get_gp_value (abfd),
fptr_sec->contents + dyn_i->fptr_offset + 8);
}
/* Return the descriptor's address. */
value = (fptr_sec->output_section->vma
+ fptr_sec->output_offset
+ dyn_i->fptr_offset);
return value;
}
/* Fill in a PLTOFF entry consisting of the function's code address
and its global pointer. Return the descriptor's address. */
static bfd_vma
set_pltoff_entry (bfd *abfd, struct bfd_link_info *info,
struct elf64_ia64_dyn_sym_info *dyn_i,
bfd_vma value, bfd_boolean is_plt)
{
struct elf64_ia64_link_hash_table *ia64_info;
asection *pltoff_sec;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return 0;
pltoff_sec = ia64_info->pltoff_sec;
/* Don't do anything if this symbol uses a real PLT entry. In
that case, we'll fill this in during finish_dynamic_symbol. */
if ((! dyn_i->want_plt || is_plt)
&& !dyn_i->pltoff_done)
{
bfd_vma gp = _bfd_get_gp_value (abfd);
/* Fill in the function descriptor. */
bfd_put_64 (abfd, value, pltoff_sec->contents + dyn_i->pltoff_offset);
bfd_put_64 (abfd, gp, pltoff_sec->contents + dyn_i->pltoff_offset + 8);
/* Install dynamic relocations if needed. */
if (!is_plt
&& info->shared
&& (!dyn_i->h
|| ELF_ST_VISIBILITY (dyn_i->h->other) == STV_DEFAULT
|| dyn_i->h->root.type != bfd_link_hash_undefweak))
{
/* VMS: */
abort ();
}
dyn_i->pltoff_done = 1;
}
/* Return the descriptor's address. */
value = (pltoff_sec->output_section->vma
+ pltoff_sec->output_offset
+ dyn_i->pltoff_offset);
return value;
}
/* Called through qsort to sort the .IA_64.unwind section during a
non-relocatable link. Set elf64_ia64_unwind_entry_compare_bfd
to the output bfd so we can do proper endianness frobbing. */
static bfd *elf64_ia64_unwind_entry_compare_bfd;
static int
elf64_ia64_unwind_entry_compare (const void * a, const void * b)
{
bfd_vma av, bv;
av = bfd_get_64 (elf64_ia64_unwind_entry_compare_bfd, a);
bv = bfd_get_64 (elf64_ia64_unwind_entry_compare_bfd, b);
return (av < bv ? -1 : av > bv ? 1 : 0);
}
/* Make sure we've got ourselves a nice fat __gp value. */
static bfd_boolean
elf64_ia64_choose_gp (bfd *abfd, struct bfd_link_info *info, bfd_boolean final)
{
bfd_vma min_vma = (bfd_vma) -1, max_vma = 0;
bfd_vma min_short_vma = min_vma, max_short_vma = 0;
struct elf_link_hash_entry *gp;
bfd_vma gp_val;
asection *os;
struct elf64_ia64_link_hash_table *ia64_info;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
/* Find the min and max vma of all sections marked short. Also collect
min and max vma of any type, for use in selecting a nice gp. */
for (os = abfd->sections; os ; os = os->next)
{
bfd_vma lo, hi;
if ((os->flags & SEC_ALLOC) == 0)
continue;
lo = os->vma;
/* When this function is called from elfNN_ia64_final_link
the correct value to use is os->size. When called from
elfNN_ia64_relax_section we are in the middle of section
sizing; some sections will already have os->size set, others
will have os->size zero and os->rawsize the previous size. */
hi = os->vma + (!final && os->rawsize ? os->rawsize : os->size);
if (hi < lo)
hi = (bfd_vma) -1;
if (min_vma > lo)
min_vma = lo;
if (max_vma < hi)
max_vma = hi;
if (os->flags & SEC_SMALL_DATA)
{
if (min_short_vma > lo)
min_short_vma = lo;
if (max_short_vma < hi)
max_short_vma = hi;
}
}
if (ia64_info->min_short_sec)
{
if (min_short_vma
> (ia64_info->min_short_sec->vma
+ ia64_info->min_short_offset))
min_short_vma = (ia64_info->min_short_sec->vma
+ ia64_info->min_short_offset);
if (max_short_vma
< (ia64_info->max_short_sec->vma
+ ia64_info->max_short_offset))
max_short_vma = (ia64_info->max_short_sec->vma
+ ia64_info->max_short_offset);
}
/* See if the user wants to force a value. */
gp = elf_link_hash_lookup (elf_hash_table (info), "__gp", FALSE,
FALSE, FALSE);
if (gp
&& (gp->root.type == bfd_link_hash_defined
|| gp->root.type == bfd_link_hash_defweak))
{
asection *gp_sec = gp->root.u.def.section;
gp_val = (gp->root.u.def.value
+ gp_sec->output_section->vma
+ gp_sec->output_offset);
}
else
{
/* Pick a sensible value. */
if (ia64_info->min_short_sec)
{
bfd_vma short_range = max_short_vma - min_short_vma;
/* If min_short_sec is set, pick one in the middle bewteen
min_short_vma and max_short_vma. */
if (short_range >= 0x400000)
goto overflow;
gp_val = min_short_vma + short_range / 2;
}
else
{
asection *got_sec = ia64_info->root.sgot;
/* Start with just the address of the .got. */
if (got_sec)
gp_val = got_sec->output_section->vma;
else if (max_short_vma != 0)
gp_val = min_short_vma;
else if (max_vma - min_vma < 0x200000)
gp_val = min_vma;
else
gp_val = max_vma - 0x200000 + 8;
}
/* If it is possible to address the entire image, but we
don't with the choice above, adjust. */
if (max_vma - min_vma < 0x400000
&& (max_vma - gp_val >= 0x200000
|| gp_val - min_vma > 0x200000))
gp_val = min_vma + 0x200000;
else if (max_short_vma != 0)
{
/* If we don't cover all the short data, adjust. */
if (max_short_vma - gp_val >= 0x200000)
gp_val = min_short_vma + 0x200000;
/* If we're addressing stuff past the end, adjust back. */
if (gp_val > max_vma)
gp_val = max_vma - 0x200000 + 8;
}
}
/* Validate whether all SHF_IA_64_SHORT sections are within
range of the chosen GP. */
if (max_short_vma != 0)
{
if (max_short_vma - min_short_vma >= 0x400000)
{
overflow:
(*_bfd_error_handler)
(_("%s: short data segment overflowed (0x%lx >= 0x400000)"),
bfd_get_filename (abfd),
(unsigned long) (max_short_vma - min_short_vma));
return FALSE;
}
else if ((gp_val > min_short_vma
&& gp_val - min_short_vma > 0x200000)
|| (gp_val < max_short_vma
&& max_short_vma - gp_val >= 0x200000))
{
(*_bfd_error_handler)
(_("%s: __gp does not cover short data segment"),
bfd_get_filename (abfd));
return FALSE;
}
}
_bfd_set_gp_value (abfd, gp_val);
return TRUE;
}
static bfd_boolean
elf64_ia64_final_link (bfd *abfd, struct bfd_link_info *info)
{
struct elf64_ia64_link_hash_table *ia64_info;
asection *unwind_output_sec;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
/* Make sure we've got ourselves a nice fat __gp value. */
if (!info->relocatable)
{
bfd_vma gp_val;
struct elf_link_hash_entry *gp;
/* We assume after gp is set, section size will only decrease. We
need to adjust gp for it. */
_bfd_set_gp_value (abfd, 0);
if (! elf64_ia64_choose_gp (abfd, info, TRUE))
return FALSE;
gp_val = _bfd_get_gp_value (abfd);
gp = elf_link_hash_lookup (elf_hash_table (info), "__gp", FALSE,
FALSE, FALSE);
if (gp)
{
gp->root.type = bfd_link_hash_defined;
gp->root.u.def.value = gp_val;
gp->root.u.def.section = bfd_abs_section_ptr;
}
}
/* If we're producing a final executable, we need to sort the contents
of the .IA_64.unwind section. Force this section to be relocated
into memory rather than written immediately to the output file. */
unwind_output_sec = NULL;
if (!info->relocatable)
{
asection *s = bfd_get_section_by_name (abfd, ELF_STRING_ia64_unwind);
if (s)
{
unwind_output_sec = s->output_section;
unwind_output_sec->contents
= bfd_malloc (unwind_output_sec->size);
if (unwind_output_sec->contents == NULL)
return FALSE;
}
}
/* Invoke the regular ELF backend linker to do all the work. */
if (!bfd_elf_final_link (abfd, info))
return FALSE;
if (unwind_output_sec)
{
elf64_ia64_unwind_entry_compare_bfd = abfd;
qsort (unwind_output_sec->contents,
(size_t) (unwind_output_sec->size / 24),
24,
elf64_ia64_unwind_entry_compare);
if (! bfd_set_section_contents (abfd, unwind_output_sec,
unwind_output_sec->contents, (bfd_vma) 0,
unwind_output_sec->size))
return FALSE;
}
return TRUE;
}
static bfd_boolean
elf64_ia64_relocate_section (bfd *output_bfd,
struct bfd_link_info *info,
bfd *input_bfd,
asection *input_section,
bfd_byte *contents,
Elf_Internal_Rela *relocs,
Elf_Internal_Sym *local_syms,
asection **local_sections)
{
struct elf64_ia64_link_hash_table *ia64_info;
Elf_Internal_Shdr *symtab_hdr;
Elf_Internal_Rela *rel;
Elf_Internal_Rela *relend;
bfd_boolean ret_val = TRUE; /* for non-fatal errors */
bfd_vma gp_val;
symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
/* Infect various flags from the input section to the output section. */
if (info->relocatable)
{
bfd_vma flags;
flags = elf_section_data(input_section)->this_hdr.sh_flags;
flags &= SHF_IA_64_NORECOV;
elf_section_data(input_section->output_section)
->this_hdr.sh_flags |= flags;
}
gp_val = _bfd_get_gp_value (output_bfd);
rel = relocs;
relend = relocs + input_section->reloc_count;
for (; rel < relend; ++rel)
{
struct elf_link_hash_entry *h;
struct elf64_ia64_dyn_sym_info *dyn_i;
bfd_reloc_status_type r;
reloc_howto_type *howto;
unsigned long r_symndx;
Elf_Internal_Sym *sym;
unsigned int r_type;
bfd_vma value;
asection *sym_sec;
bfd_byte *hit_addr;
bfd_boolean dynamic_symbol_p;
bfd_boolean undef_weak_ref;
r_type = ELF64_R_TYPE (rel->r_info);
if (r_type > R_IA64_MAX_RELOC_CODE)
{
(*_bfd_error_handler)
(_("%B: unknown relocation type %d"),
input_bfd, (int) r_type);
bfd_set_error (bfd_error_bad_value);
ret_val = FALSE;
continue;
}
howto = ia64_elf_lookup_howto (r_type);
r_symndx = ELF64_R_SYM (rel->r_info);
h = NULL;
sym = NULL;
sym_sec = NULL;
undef_weak_ref = FALSE;
if (r_symndx < symtab_hdr->sh_info)
{
/* Reloc against local symbol. */
asection *msec;
sym = local_syms + r_symndx;
sym_sec = local_sections[r_symndx];
msec = sym_sec;
value = _bfd_elf_rela_local_sym (output_bfd, sym, &msec, rel);
if (!info->relocatable
&& (sym_sec->flags & SEC_MERGE) != 0
&& ELF_ST_TYPE (sym->st_info) == STT_SECTION
&& sym_sec->sec_info_type == SEC_INFO_TYPE_MERGE)
{
struct elf64_ia64_local_hash_entry *loc_h;
loc_h = get_local_sym_hash (ia64_info, input_bfd, rel, FALSE);
if (loc_h && ! loc_h->sec_merge_done)
{
struct elf64_ia64_dyn_sym_info *dynent;
unsigned int count;
for (count = loc_h->count, dynent = loc_h->info;
count != 0;
count--, dynent++)
{
msec = sym_sec;
dynent->addend =
_bfd_merged_section_offset (output_bfd, &msec,
elf_section_data (msec)->
sec_info,
sym->st_value
+ dynent->addend);
dynent->addend -= sym->st_value;
dynent->addend += msec->output_section->vma
+ msec->output_offset
- sym_sec->output_section->vma
- sym_sec->output_offset;
}
/* We may have introduced duplicated entries. We need
to remove them properly. */
count = sort_dyn_sym_info (loc_h->info, loc_h->count);
if (count != loc_h->count)
{
loc_h->count = count;
loc_h->sorted_count = count;
}
loc_h->sec_merge_done = 1;
}
}
}
else
{
bfd_boolean unresolved_reloc;
bfd_boolean warned;
struct elf_link_hash_entry **sym_hashes = elf_sym_hashes (input_bfd);
RELOC_FOR_GLOBAL_SYMBOL (info, input_bfd, input_section, rel,
r_symndx, symtab_hdr, sym_hashes,
h, sym_sec, value,
unresolved_reloc, warned);
if (h->root.type == bfd_link_hash_undefweak)
undef_weak_ref = TRUE;
else if (warned)
continue;
}
/* For relocs against symbols from removed linkonce sections,
or sections discarded by a linker script, we just want the
section contents zeroed. Avoid any special processing. */
if (sym_sec != NULL && discarded_section (sym_sec))
RELOC_AGAINST_DISCARDED_SECTION (info, input_bfd, input_section,
rel, 1, relend, howto, 0, contents);
if (info->relocatable)
continue;
hit_addr = contents + rel->r_offset;
value += rel->r_addend;
dynamic_symbol_p = elf64_ia64_dynamic_symbol_p (h);
switch (r_type)
{
case R_IA64_NONE:
case R_IA64_LDXMOV:
continue;
case R_IA64_IMM14:
case R_IA64_IMM22:
case R_IA64_IMM64:
case R_IA64_DIR32MSB:
case R_IA64_DIR32LSB:
case R_IA64_DIR64MSB:
case R_IA64_DIR64LSB:
/* Install a dynamic relocation for this reloc. */
if ((dynamic_symbol_p || info->shared)
&& r_symndx != 0
&& (input_section->flags & SEC_ALLOC) != 0)
{
unsigned int dyn_r_type;
bfd_vma addend;
switch (r_type)
{
case R_IA64_IMM14:
case R_IA64_IMM22:
case R_IA64_IMM64:
/* ??? People shouldn't be doing non-pic code in
shared libraries nor dynamic executables. */
(*_bfd_error_handler)
(_("%B: non-pic code with imm relocation against dynamic symbol `%s'"),
input_bfd,
h ? h->root.root.string
: bfd_elf_sym_name (input_bfd, symtab_hdr, sym,
sym_sec));
ret_val = FALSE;
continue;
default:
break;
}
/* If we don't need dynamic symbol lookup, find a
matching RELATIVE relocation. */
dyn_r_type = r_type;
if (dynamic_symbol_p)
{
addend = rel->r_addend;
value = 0;
}
else
{
addend = value;
}
/* VMS: install a FIX64. */
switch (dyn_r_type)
{
case R_IA64_DIR32LSB:
dyn_r_type = R_IA64_VMS_FIX32;
break;
case R_IA64_DIR64LSB:
dyn_r_type = R_IA64_VMS_FIX64;
break;
default:
BFD_ASSERT (FALSE);
break;
}
elf64_ia64_install_fixup
(output_bfd, ia64_info, h,
dyn_r_type, input_section, rel->r_offset, addend);
r = bfd_reloc_ok;
break;
}
/* Fall through. */
case R_IA64_LTV32MSB:
case R_IA64_LTV32LSB:
case R_IA64_LTV64MSB:
case R_IA64_LTV64LSB:
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_GPREL22:
case R_IA64_GPREL64I:
case R_IA64_GPREL32MSB:
case R_IA64_GPREL32LSB:
case R_IA64_GPREL64MSB:
case R_IA64_GPREL64LSB:
if (dynamic_symbol_p)
{
(*_bfd_error_handler)
(_("%B: @gprel relocation against dynamic symbol %s"),
input_bfd,
h ? h->root.root.string
: bfd_elf_sym_name (input_bfd, symtab_hdr, sym,
sym_sec));
ret_val = FALSE;
continue;
}
value -= gp_val;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_LTOFF22:
case R_IA64_LTOFF22X:
case R_IA64_LTOFF64I:
dyn_i = get_dyn_sym_info (ia64_info, h, input_bfd, rel, FALSE);
value = set_got_entry (input_bfd, info, dyn_i,
rel->r_addend, value, R_IA64_DIR64LSB);
value -= gp_val;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_PLTOFF22:
case R_IA64_PLTOFF64I:
case R_IA64_PLTOFF64MSB:
case R_IA64_PLTOFF64LSB:
dyn_i = get_dyn_sym_info (ia64_info, h, input_bfd, rel, FALSE);
value = set_pltoff_entry (output_bfd, info, dyn_i, value, FALSE);
value -= gp_val;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_FPTR64I:
case R_IA64_FPTR32MSB:
case R_IA64_FPTR32LSB:
case R_IA64_FPTR64MSB:
case R_IA64_FPTR64LSB:
dyn_i = get_dyn_sym_info (ia64_info, h, input_bfd, rel, FALSE);
if (dyn_i->want_fptr)
{
if (!undef_weak_ref)
value = set_fptr_entry (output_bfd, info, dyn_i, value);
}
if (!dyn_i->want_fptr || info->pie)
{
/* Otherwise, we expect the dynamic linker to create
the entry. */
if (dyn_i->want_fptr)
{
if (r_type == R_IA64_FPTR64I)
{
/* We can't represent this without a dynamic symbol.
Adjust the relocation to be against an output
section symbol, which are always present in the
dynamic symbol table. */
/* ??? People shouldn't be doing non-pic code in
shared libraries. Hork. */
(*_bfd_error_handler)
(_("%B: linking non-pic code in a position independent executable"),
input_bfd);
ret_val = FALSE;
continue;
}
}
else
{
value = 0;
}
/* VMS: FIXFD. */
elf64_ia64_install_fixup
(output_bfd, ia64_info, h, R_IA64_VMS_FIXFD,
input_section, rel->r_offset, 0);
r = bfd_reloc_ok;
break;
}
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_LTOFF_FPTR22:
case R_IA64_LTOFF_FPTR64I:
case R_IA64_LTOFF_FPTR32MSB:
case R_IA64_LTOFF_FPTR32LSB:
case R_IA64_LTOFF_FPTR64MSB:
case R_IA64_LTOFF_FPTR64LSB:
dyn_i = get_dyn_sym_info (ia64_info, h, input_bfd, rel, FALSE);
if (dyn_i->want_fptr)
{
BFD_ASSERT (h == NULL || !h->def_dynamic);
if (!undef_weak_ref)
value = set_fptr_entry (output_bfd, info, dyn_i, value);
}
else
value = 0;
value = set_got_entry (output_bfd, info, dyn_i,
rel->r_addend, value, R_IA64_FPTR64LSB);
value -= gp_val;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_PCREL32MSB:
case R_IA64_PCREL32LSB:
case R_IA64_PCREL64MSB:
case R_IA64_PCREL64LSB:
/* Install a dynamic relocation for this reloc. */
if (dynamic_symbol_p && r_symndx != 0)
{
/* VMS: doesn't exist ??? */
abort ();
}
goto finish_pcrel;
case R_IA64_PCREL21B:
case R_IA64_PCREL60B:
/* We should have created a PLT entry for any dynamic symbol. */
dyn_i = NULL;
if (h)
dyn_i = get_dyn_sym_info (ia64_info, h, NULL, NULL, FALSE);
if (dyn_i && dyn_i->want_plt2)
{
/* Should have caught this earlier. */
BFD_ASSERT (rel->r_addend == 0);
value = (ia64_info->root.splt->output_section->vma
+ ia64_info->root.splt->output_offset
+ dyn_i->plt2_offset);
}
else
{
/* Since there's no PLT entry, Validate that this is
locally defined. */
BFD_ASSERT (undef_weak_ref || sym_sec->output_section != NULL);
/* If the symbol is undef_weak, we shouldn't be trying
to call it. There's every chance that we'd wind up
with an out-of-range fixup here. Don't bother setting
any value at all. */
if (undef_weak_ref)
continue;
}
goto finish_pcrel;
case R_IA64_PCREL21BI:
case R_IA64_PCREL21F:
case R_IA64_PCREL21M:
case R_IA64_PCREL22:
case R_IA64_PCREL64I:
/* The PCREL21BI reloc is specifically not intended for use with
dynamic relocs. PCREL21F and PCREL21M are used for speculation
fixup code, and thus probably ought not be dynamic. The
PCREL22 and PCREL64I relocs aren't emitted as dynamic relocs. */
if (dynamic_symbol_p)
{
const char *msg;
if (r_type == R_IA64_PCREL21BI)
msg = _("%B: @internal branch to dynamic symbol %s");
else if (r_type == R_IA64_PCREL21F || r_type == R_IA64_PCREL21M)
msg = _("%B: speculation fixup to dynamic symbol %s");
else
msg = _("%B: @pcrel relocation against dynamic symbol %s");
(*_bfd_error_handler) (msg, input_bfd,
h ? h->root.root.string
: bfd_elf_sym_name (input_bfd,
symtab_hdr,
sym,
sym_sec));
ret_val = FALSE;
continue;
}
goto finish_pcrel;
finish_pcrel:
/* Make pc-relative. */
value -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset) & ~ (bfd_vma) 0x3;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_SEGREL32MSB:
case R_IA64_SEGREL32LSB:
case R_IA64_SEGREL64MSB:
case R_IA64_SEGREL64LSB:
{
/* Find the segment that contains the output_section. */
Elf_Internal_Phdr *p = _bfd_elf_find_segment_containing_section
(output_bfd, sym_sec->output_section);
if (p == NULL)
{
r = bfd_reloc_notsupported;
}
else
{
/* The VMA of the segment is the vaddr of the associated
program header. */
if (value > p->p_vaddr)
value -= p->p_vaddr;
else
value = 0;
r = ia64_elf_install_value (hit_addr, value, r_type);
}
break;
}
case R_IA64_SECREL32MSB:
case R_IA64_SECREL32LSB:
case R_IA64_SECREL64MSB:
case R_IA64_SECREL64LSB:
/* Make output-section relative to section where the symbol
is defined. PR 475 */
if (sym_sec)
value -= sym_sec->output_section->vma;
r = ia64_elf_install_value (hit_addr, value, r_type);
break;
case R_IA64_IPLTMSB:
case R_IA64_IPLTLSB:
/* Install a dynamic relocation for this reloc. */
if ((dynamic_symbol_p || info->shared)
&& (input_section->flags & SEC_ALLOC) != 0)
{
/* VMS: FIXFD ?? */
abort ();
}
if (r_type == R_IA64_IPLTMSB)
r_type = R_IA64_DIR64MSB;
else
r_type = R_IA64_DIR64LSB;
ia64_elf_install_value (hit_addr, value, r_type);
r = ia64_elf_install_value (hit_addr + 8, gp_val, r_type);
break;
case R_IA64_TPREL14:
case R_IA64_TPREL22:
case R_IA64_TPREL64I:
r = bfd_reloc_notsupported;
break;
case R_IA64_DTPREL14:
case R_IA64_DTPREL22:
case R_IA64_DTPREL64I:
case R_IA64_DTPREL32LSB:
case R_IA64_DTPREL32MSB:
case R_IA64_DTPREL64LSB:
case R_IA64_DTPREL64MSB:
r = bfd_reloc_notsupported;
break;
case R_IA64_LTOFF_TPREL22:
case R_IA64_LTOFF_DTPMOD22:
case R_IA64_LTOFF_DTPREL22:
r = bfd_reloc_notsupported;
break;
default:
r = bfd_reloc_notsupported;
break;
}
switch (r)
{
case bfd_reloc_ok:
break;
case bfd_reloc_undefined:
/* This can happen for global table relative relocs if
__gp is undefined. This is a panic situation so we
don't try to continue. */
(*info->callbacks->undefined_symbol)
(info, "__gp", input_bfd, input_section, rel->r_offset, 1);
return FALSE;
case bfd_reloc_notsupported:
{
const char *name;
if (h)
name = h->root.root.string;
else
name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym,
sym_sec);
if (!(*info->callbacks->warning) (info, _("unsupported reloc"),
name, input_bfd,
input_section, rel->r_offset))
return FALSE;
ret_val = FALSE;
}
break;
case bfd_reloc_dangerous:
case bfd_reloc_outofrange:
case bfd_reloc_overflow:
default:
{
const char *name;
if (h)
name = h->root.root.string;
else
name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym,
sym_sec);
switch (r_type)
{
case R_IA64_TPREL14:
case R_IA64_TPREL22:
case R_IA64_TPREL64I:
case R_IA64_DTPREL14:
case R_IA64_DTPREL22:
case R_IA64_DTPREL64I:
case R_IA64_DTPREL32LSB:
case R_IA64_DTPREL32MSB:
case R_IA64_DTPREL64LSB:
case R_IA64_DTPREL64MSB:
case R_IA64_LTOFF_TPREL22:
case R_IA64_LTOFF_DTPMOD22:
case R_IA64_LTOFF_DTPREL22:
(*_bfd_error_handler)
(_("%B: missing TLS section for relocation %s against `%s' at 0x%lx in section `%A'."),
input_bfd, input_section, howto->name, name,
rel->r_offset);
break;
case R_IA64_PCREL21B:
case R_IA64_PCREL21BI:
case R_IA64_PCREL21M:
case R_IA64_PCREL21F:
if (is_elf_hash_table (info->hash))
{
/* Relaxtion is always performed for ELF output.
Overflow failures for those relocations mean
that the section is too big to relax. */
(*_bfd_error_handler)
(_("%B: Can't relax br (%s) to `%s' at 0x%lx in section `%A' with size 0x%lx (> 0x1000000)."),
input_bfd, input_section, howto->name, name,
rel->r_offset, input_section->size);
break;
}
default:
if (!(*info->callbacks->reloc_overflow) (info,
&h->root,
name,
howto->name,
(bfd_vma) 0,
input_bfd,
input_section,
rel->r_offset))
return FALSE;
break;
}
ret_val = FALSE;
}
break;
}
}
return ret_val;
}
static bfd_boolean
elf64_ia64_finish_dynamic_symbol (bfd *output_bfd,
struct bfd_link_info *info,
struct elf_link_hash_entry *h,
Elf_Internal_Sym *sym)
{
struct elf64_ia64_link_hash_table *ia64_info;
struct elf64_ia64_dyn_sym_info *dyn_i;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
dyn_i = get_dyn_sym_info (ia64_info, h, NULL, NULL, FALSE);
/* Fill in the PLT data, if required. */
if (dyn_i && dyn_i->want_plt)
{
bfd_byte *loc;
asection *plt_sec;
bfd_vma plt_addr, pltoff_addr, gp_val;
gp_val = _bfd_get_gp_value (output_bfd);
plt_sec = ia64_info->root.splt;
plt_addr = 0; /* Not used as overriden by FIXUPs. */
pltoff_addr = set_pltoff_entry (output_bfd, info, dyn_i, plt_addr, TRUE);
/* Initialize the FULL PLT entry, if needed. */
if (dyn_i->want_plt2)
{
loc = plt_sec->contents + dyn_i->plt2_offset;
memcpy (loc, plt_full_entry, PLT_FULL_ENTRY_SIZE);
ia64_elf_install_value (loc, pltoff_addr - gp_val, R_IA64_IMM22);
/* Mark the symbol as undefined, rather than as defined in the
plt section. Leave the value alone. */
/* ??? We didn't redefine it in adjust_dynamic_symbol in the
first place. But perhaps elflink.c did some for us. */
if (!h->def_regular)
sym->st_shndx = SHN_UNDEF;
}
/* VMS: FIXFD. */
elf64_ia64_install_fixup
(output_bfd, ia64_info, h, R_IA64_VMS_FIXFD, ia64_info->pltoff_sec,
pltoff_addr - (ia64_info->pltoff_sec->output_section->vma
+ ia64_info->pltoff_sec->output_offset), 0);
}
/* Mark some specially defined symbols as absolute. */
if (h == ia64_info->root.hdynamic
|| h == ia64_info->root.hgot
|| h == ia64_info->root.hplt)
sym->st_shndx = SHN_ABS;
return TRUE;
}
static bfd_boolean
elf64_ia64_finish_dynamic_sections (bfd *abfd,
struct bfd_link_info *info)
{
struct elf64_ia64_link_hash_table *ia64_info;
bfd *dynobj;
ia64_info = elf64_ia64_hash_table (info);
if (ia64_info == NULL)
return FALSE;
dynobj = ia64_info->root.dynobj;
if (elf_hash_table (info)->dynamic_sections_created)
{
Elf64_External_Dyn *dyncon, *dynconend;
asection *sdyn;
asection *unwind_sec;
bfd_vma gp_val;
unsigned int gp_seg;
bfd_vma gp_off;
Elf_Internal_Phdr *phdr;
Elf_Internal_Phdr *base_phdr;
unsigned int unwind_seg = 0;
unsigned int code_seg = 0;
sdyn = bfd_get_linker_section (dynobj, ".dynamic");
BFD_ASSERT (sdyn != NULL);
dyncon = (Elf64_External_Dyn *) sdyn->contents;
dynconend = (Elf64_External_Dyn *) (sdyn->contents + sdyn->size);
gp_val = _bfd_get_gp_value (abfd);
phdr = _bfd_elf_find_segment_containing_section
(info->output_bfd, ia64_info->pltoff_sec->output_section);
BFD_ASSERT (phdr != NULL);
base_phdr = elf_tdata (info->output_bfd)->phdr;
gp_seg = phdr - base_phdr;
gp_off = gp_val - phdr->p_vaddr;
unwind_sec = bfd_get_section_by_name (abfd, ELF_STRING_ia64_unwind);
if (unwind_sec != NULL)
{
asection *code_sec;
phdr = _bfd_elf_find_segment_containing_section (abfd, unwind_sec);
BFD_ASSERT (phdr != NULL);
unwind_seg = phdr - base_phdr;
code_sec = bfd_get_section_by_name (abfd, "$CODE$");
phdr = _bfd_elf_find_segment_containing_section (abfd, code_sec);
BFD_ASSERT (phdr != NULL);
code_seg = phdr - base_phdr;
}
for (; dyncon < dynconend; dyncon++)
{
Elf_Internal_Dyn dyn;
bfd_elf64_swap_dyn_in (dynobj, dyncon, &dyn);
switch (dyn.d_tag)
{
case DT_IA_64_VMS_FIXUP_RELA_OFF:
dyn.d_un.d_val +=
(ia64_info->fixups_sec->output_section->vma
+ ia64_info->fixups_sec->output_offset)
- (sdyn->output_section->vma + sdyn->output_offset);
break;
case DT_IA_64_VMS_PLTGOT_OFFSET:
dyn.d_un.d_val = gp_off;
break;
case DT_IA_64_VMS_PLTGOT_SEG:
dyn.d_un.d_val = gp_seg;
break;
case DT_IA_64_VMS_UNWINDSZ:
if (unwind_sec == NULL)
{
dyn.d_tag = DT_NULL;
dyn.d_un.d_val = 0xdead;
}
else
dyn.d_un.d_val = unwind_sec->size;
break;
case DT_IA_64_VMS_UNWIND_CODSEG:
dyn.d_un.d_val = code_seg;
break;
case DT_IA_64_VMS_UNWIND_INFOSEG:
case DT_IA_64_VMS_UNWIND_SEG:
dyn.d_un.d_val = unwind_seg;
break;
case DT_IA_64_VMS_UNWIND_OFFSET:
break;
default:
/* No need to rewrite the entry. */
continue;
}
bfd_elf64_swap_dyn_out (abfd, &dyn, dyncon);
}
}
/* Handle transfer addresses. */
{
asection *tfr_sec = ia64_info->transfer_sec;
struct elf64_vms_transfer *tfr;
struct elf_link_hash_entry *tfr3;
tfr = (struct elf64_vms_transfer *)tfr_sec->contents;
bfd_putl32 (6 * 8, tfr->size);
bfd_putl64 (tfr_sec->output_section->vma
+ tfr_sec->output_offset
+ 6 * 8, tfr->tfradr3);
tfr3 = elf_link_hash_lookup (elf_hash_table (info), "ELF$TFRADR", FALSE,
FALSE, FALSE);
if (tfr3
&& (tfr3->root.type == bfd_link_hash_defined
|| tfr3->root.type == bfd_link_hash_defweak))
{
asection *tfr3_sec = tfr3->root.u.def.section;
bfd_vma tfr3_val;
tfr3_val = (tfr3->root.u.def.value
+ tfr3_sec->output_section->vma
+ tfr3_sec->output_offset);
bfd_putl64 (tfr3_val, tfr->tfr3_func);
bfd_putl64 (_bfd_get_gp_value (info->output_bfd), tfr->tfr3_gp);
}
/* FIXME: set linker flags,
handle lib$initialize. */
}
return TRUE;
}
/* ELF file flag handling: */
/* Function to keep IA-64 specific file flags. */
static bfd_boolean
elf64_ia64_set_private_flags (bfd *abfd, flagword flags)
{
BFD_ASSERT (!elf_flags_init (abfd)
|| elf_elfheader (abfd)->e_flags == flags);
elf_elfheader (abfd)->e_flags = flags;
elf_flags_init (abfd) = TRUE;
return TRUE;
}
/* Merge backend specific data from an object file to the output
object file when linking. */
static bfd_boolean
elf64_ia64_merge_private_bfd_data (bfd *ibfd, bfd *obfd)
{
flagword out_flags;
flagword in_flags;
bfd_boolean ok = TRUE;
/* Don't even pretend to support mixed-format linking. */
if (bfd_get_flavour (ibfd) != bfd_target_elf_flavour
|| bfd_get_flavour (obfd) != bfd_target_elf_flavour)
return FALSE;
in_flags = elf_elfheader (ibfd)->e_flags;
out_flags = elf_elfheader (obfd)->e_flags;
if (! elf_flags_init (obfd))
{
elf_flags_init (obfd) = TRUE;
elf_elfheader (obfd)->e_flags = in_flags;
if (bfd_get_arch (obfd) == bfd_get_arch (ibfd)
&& bfd_get_arch_info (obfd)->the_default)
{
return bfd_set_arch_mach (obfd, bfd_get_arch (ibfd),
bfd_get_mach (ibfd));
}
return TRUE;
}
/* Check flag compatibility. */
if (in_flags == out_flags)
return TRUE;
/* Output has EF_IA_64_REDUCEDFP set only if all inputs have it set. */
if (!(in_flags & EF_IA_64_REDUCEDFP) && (out_flags & EF_IA_64_REDUCEDFP))
elf_elfheader (obfd)->e_flags &= ~EF_IA_64_REDUCEDFP;
if ((in_flags & EF_IA_64_TRAPNIL) != (out_flags & EF_IA_64_TRAPNIL))
{
(*_bfd_error_handler)
(_("%B: linking trap-on-NULL-dereference with non-trapping files"),
ibfd);
bfd_set_error (bfd_error_bad_value);
ok = FALSE;
}
if ((in_flags & EF_IA_64_BE) != (out_flags & EF_IA_64_BE))
{
(*_bfd_error_handler)
(_("%B: linking big-endian files with little-endian files"),
ibfd);
bfd_set_error (bfd_error_bad_value);
ok = FALSE;
}
if ((in_flags & EF_IA_64_ABI64) != (out_flags & EF_IA_64_ABI64))
{
(*_bfd_error_handler)
(_("%B: linking 64-bit files with 32-bit files"),
ibfd);
bfd_set_error (bfd_error_bad_value);
ok = FALSE;
}
if ((in_flags & EF_IA_64_CONS_GP) != (out_flags & EF_IA_64_CONS_GP))
{
(*_bfd_error_handler)
(_("%B: linking constant-gp files with non-constant-gp files"),
ibfd);
bfd_set_error (bfd_error_bad_value);
ok = FALSE;
}
if ((in_flags & EF_IA_64_NOFUNCDESC_CONS_GP)
!= (out_flags & EF_IA_64_NOFUNCDESC_CONS_GP))
{
(*_bfd_error_handler)
(_("%B: linking auto-pic files with non-auto-pic files"),
ibfd);
bfd_set_error (bfd_error_bad_value);
ok = FALSE;
}
return ok;
}
static bfd_boolean
elf64_ia64_print_private_bfd_data (bfd *abfd, void * ptr)
{
FILE *file = (FILE *) ptr;
flagword flags = elf_elfheader (abfd)->e_flags;
BFD_ASSERT (abfd != NULL && ptr != NULL);
fprintf (file, "private flags = %s%s%s%s%s%s%s%s\n",
(flags & EF_IA_64_TRAPNIL) ? "TRAPNIL, " : "",
(flags & EF_IA_64_EXT) ? "EXT, " : "",
(flags & EF_IA_64_BE) ? "BE, " : "LE, ",
(flags & EF_IA_64_REDUCEDFP) ? "REDUCEDFP, " : "",
(flags & EF_IA_64_CONS_GP) ? "CONS_GP, " : "",
(flags & EF_IA_64_NOFUNCDESC_CONS_GP) ? "NOFUNCDESC_CONS_GP, " : "",
(flags & EF_IA_64_ABSOLUTE) ? "ABSOLUTE, " : "",
(flags & EF_IA_64_ABI64) ? "ABI64" : "ABI32");
_bfd_elf_print_private_bfd_data (abfd, ptr);
return TRUE;
}
static enum elf_reloc_type_class
elf64_ia64_reloc_type_class (const struct bfd_link_info *info ATTRIBUTE_UNUSED,
const asection *rel_sec ATTRIBUTE_UNUSED,
const Elf_Internal_Rela *rela)
{
switch ((int) ELF64_R_TYPE (rela->r_info))
{
case R_IA64_REL32MSB:
case R_IA64_REL32LSB:
case R_IA64_REL64MSB:
case R_IA64_REL64LSB:
return reloc_class_relative;
case R_IA64_IPLTMSB:
case R_IA64_IPLTLSB:
return reloc_class_plt;
case R_IA64_COPY:
return reloc_class_copy;
default:
return reloc_class_normal;
}
}
static const struct bfd_elf_special_section elf64_ia64_special_sections[] =
{
{ STRING_COMMA_LEN (".sbss"), -1, SHT_NOBITS, SHF_ALLOC + SHF_WRITE + SHF_IA_64_SHORT },
{ STRING_COMMA_LEN (".sdata"), -1, SHT_PROGBITS, SHF_ALLOC + SHF_WRITE + SHF_IA_64_SHORT },
{ NULL, 0, 0, 0, 0 }
};
static bfd_boolean
elf64_ia64_object_p (bfd *abfd)
{
asection *sec;
asection *group, *unwi, *unw;
flagword flags;
const char *name;
char *unwi_name, *unw_name;
bfd_size_type amt;
if (abfd->flags & DYNAMIC)
return TRUE;
/* Flags for fake group section. */
flags = (SEC_LINKER_CREATED | SEC_GROUP | SEC_LINK_ONCE
| SEC_EXCLUDE);
/* We add a fake section group for each .gnu.linkonce.t.* section,
which isn't in a section group, and its unwind sections. */
for (sec = abfd->sections; sec != NULL; sec = sec->next)
{
if (elf_sec_group (sec) == NULL
&& ((sec->flags & (SEC_LINK_ONCE | SEC_CODE | SEC_GROUP))
== (SEC_LINK_ONCE | SEC_CODE))
&& CONST_STRNEQ (sec->name, ".gnu.linkonce.t."))
{
name = sec->name + 16;
amt = strlen (name) + sizeof (".gnu.linkonce.ia64unwi.");
unwi_name = bfd_alloc (abfd, amt);
if (!unwi_name)
return FALSE;
strcpy (stpcpy (unwi_name, ".gnu.linkonce.ia64unwi."), name);
unwi = bfd_get_section_by_name (abfd, unwi_name);
amt = strlen (name) + sizeof (".gnu.linkonce.ia64unw.");
unw_name = bfd_alloc (abfd, amt);
if (!unw_name)
return FALSE;
strcpy (stpcpy (unw_name, ".gnu.linkonce.ia64unw."), name);
unw = bfd_get_section_by_name (abfd, unw_name);
/* We need to create a fake group section for it and its
unwind sections. */
group = bfd_make_section_anyway_with_flags (abfd, name,
flags);
if (group == NULL)
return FALSE;
/* Move the fake group section to the beginning. */
bfd_section_list_remove (abfd, group);
bfd_section_list_prepend (abfd, group);
elf_next_in_group (group) = sec;
elf_group_name (sec) = name;
elf_next_in_group (sec) = sec;
elf_sec_group (sec) = group;
if (unwi)
{
elf_group_name (unwi) = name;
elf_next_in_group (unwi) = sec;
elf_next_in_group (sec) = unwi;
elf_sec_group (unwi) = group;
}
if (unw)
{
elf_group_name (unw) = name;
if (unwi)
{
elf_next_in_group (unw) = elf_next_in_group (unwi);
elf_next_in_group (unwi) = unw;
}
else
{
elf_next_in_group (unw) = sec;
elf_next_in_group (sec) = unw;
}
elf_sec_group (unw) = group;
}
/* Fake SHT_GROUP section header. */
elf_section_data (group)->this_hdr.bfd_section = group;
elf_section_data (group)->this_hdr.sh_type = SHT_GROUP;
}
}
return TRUE;
}
/* Handle an IA-64 specific section when reading an object file. This
is called when bfd_section_from_shdr finds a section with an unknown
type. */
static bfd_boolean
elf64_vms_section_from_shdr (bfd *abfd,
Elf_Internal_Shdr *hdr,
const char *name,
int shindex)
{
flagword secflags = 0;
switch (hdr->sh_type)
{
case SHT_IA_64_VMS_TRACE:
case SHT_IA_64_VMS_DEBUG:
case SHT_IA_64_VMS_DEBUG_STR:
secflags = SEC_DEBUGGING;
break;
case SHT_IA_64_UNWIND:
case SHT_IA_64_HP_OPT_ANOT:
break;
case SHT_IA_64_EXT:
if (strcmp (name, ELF_STRING_ia64_archext) != 0)
return FALSE;
break;
default:
return FALSE;
}
if (! _bfd_elf_make_section_from_shdr (abfd, hdr, name, shindex))
return FALSE;
if (secflags != 0)
{
asection *newsect = hdr->bfd_section;
if (! bfd_set_section_flags
(abfd, newsect, bfd_get_section_flags (abfd, newsect) | secflags))
return FALSE;
}
return TRUE;
}
static bfd_boolean
elf64_vms_object_p (bfd *abfd)
{
Elf_Internal_Ehdr *i_ehdrp = elf_elfheader (abfd);
Elf_Internal_Phdr *i_phdr = elf_tdata (abfd)->phdr;
unsigned int i;
unsigned int num_text = 0;
unsigned int num_data = 0;
unsigned int num_rodata = 0;
char name[16];
if (!elf64_ia64_object_p (abfd))
return FALSE;
/* Many VMS compilers do not generate sections for the corresponding
segment. This is boring as binutils tools won't be able to disassemble
the code. So we simply create all the missing sections. */
for (i = 0; i < i_ehdrp->e_phnum; i++, i_phdr++)
{
/* Is there a section for this segment? */
bfd_vma base_vma = i_phdr->p_vaddr;
bfd_vma limit_vma = base_vma + i_phdr->p_filesz;
if (i_phdr->p_type != PT_LOAD)
continue;
/* We need to cover from base_vms to limit_vma. */
again:
while (base_vma < limit_vma)
{
bfd_vma next_vma = limit_vma;
asection *nsec;
asection *sec;
flagword flags;
char *nname = NULL;
/* Find a section covering [base_vma;limit_vma) */
for (sec = abfd->sections; sec != NULL; sec = sec->next)
{
/* Skip uninteresting sections (either not in memory or
below base_vma. */
if ((sec->flags & (SEC_ALLOC | SEC_LOAD)) == 0
|| sec->vma + sec->size <= base_vma)
continue;
if (sec->vma <= base_vma)
{
/* This section covers (maybe partially) the beginning
of the range. */
base_vma = sec->vma + sec->size;
goto again;
}
if (sec->vma < next_vma)
{
/* This section partially covers the end of the range.
Used to compute the size of the hole. */
next_vma = sec->vma;
}
}
/* No section covering [base_vma; next_vma). Create a fake one. */
flags = SEC_ALLOC | SEC_LOAD | SEC_HAS_CONTENTS;
if (i_phdr->p_flags & PF_X)
{
flags |= SEC_CODE;
if (num_text++ == 0)
nname = ".text";
else
sprintf (name, ".text$%u", num_text);
}
else if ((i_phdr->p_flags & (PF_R | PF_W)) == PF_R)
{
flags |= SEC_READONLY;
sprintf (name, ".rodata$%u", num_rodata++);
}
else
{
flags |= SEC_DATA;
sprintf (name, ".data$%u", num_data++);
}
/* Allocate name. */
if (nname == NULL)
{
size_t name_len = strlen (name) + 1;
nname = bfd_alloc (abfd, name_len);
if (nname == NULL)
return FALSE;
memcpy (nname, name, name_len);
}
/* Create and fill new section. */
nsec = bfd_make_section_anyway_with_flags (abfd, nname, flags);
if (nsec == NULL)
return FALSE;
nsec->vma = base_vma;
nsec->size = next_vma - base_vma;
nsec->filepos = i_phdr->p_offset + (base_vma - i_phdr->p_vaddr);
base_vma = next_vma;
}
}
return TRUE;
}
static void
elf64_vms_post_process_headers (bfd *abfd,
struct bfd_link_info *info ATTRIBUTE_UNUSED)
{
Elf_Internal_Ehdr *i_ehdrp = elf_elfheader (abfd);
i_ehdrp->e_ident[EI_OSABI] = ELFOSABI_OPENVMS;
i_ehdrp->e_ident[EI_ABIVERSION] = 2;
}
static bfd_boolean
elf64_vms_section_processing (bfd *abfd ATTRIBUTE_UNUSED,
Elf_Internal_Shdr *hdr)
{
if (hdr->bfd_section != NULL)
{
const char *name = bfd_get_section_name (abfd, hdr->bfd_section);
if (strcmp (name, ".text") == 0)
hdr->sh_flags |= SHF_IA_64_VMS_SHARED;
else if ((strcmp (name, ".debug") == 0)
|| (strcmp (name, ".debug_abbrev") == 0)
|| (strcmp (name, ".debug_aranges") == 0)
|| (strcmp (name, ".debug_frame") == 0)
|| (strcmp (name, ".debug_info") == 0)
|| (strcmp (name, ".debug_loc") == 0)
|| (strcmp (name, ".debug_macinfo") == 0)
|| (strcmp (name, ".debug_pubnames") == 0)
|| (strcmp (name, ".debug_pubtypes") == 0))
hdr->sh_type = SHT_IA_64_VMS_DEBUG;
else if ((strcmp (name, ".debug_line") == 0)
|| (strcmp (name, ".debug_ranges") == 0)
|| (strcmp (name, ".trace_info") == 0)
|| (strcmp (name, ".trace_abbrev") == 0)
|| (strcmp (name, ".trace_aranges") == 0))
hdr->sh_type = SHT_IA_64_VMS_TRACE;
else if (strcmp (name, ".debug_str") == 0)
hdr->sh_type = SHT_IA_64_VMS_DEBUG_STR;
}
return TRUE;
}
/* The final processing done just before writing out a VMS IA-64 ELF
object file. */
static void
elf64_vms_final_write_processing (bfd *abfd,
bfd_boolean linker ATTRIBUTE_UNUSED)
{
Elf_Internal_Shdr *hdr;
asection *s;
int unwind_info_sect_idx = 0;
for (s = abfd->sections; s; s = s->next)
{
hdr = &elf_section_data (s)->this_hdr;
if (strcmp (bfd_get_section_name (abfd, hdr->bfd_section),
".IA_64.unwind_info") == 0)
unwind_info_sect_idx = elf_section_data (s)->this_idx;
switch (hdr->sh_type)
{
case SHT_IA_64_UNWIND:
/* VMS requires sh_info to point to the unwind info section. */
hdr->sh_info = unwind_info_sect_idx;
break;
}
}
if (! elf_flags_init (abfd))
{
unsigned long flags = 0;
if (abfd->xvec->byteorder == BFD_ENDIAN_BIG)
flags |= EF_IA_64_BE;
if (bfd_get_mach (abfd) == bfd_mach_ia64_elf64)
flags |= EF_IA_64_ABI64;
elf_elfheader (abfd)->e_flags = flags;
elf_flags_init (abfd) = TRUE;
}
}
static bfd_boolean
elf64_vms_write_shdrs_and_ehdr (bfd *abfd)
{
unsigned char needed_count[8];
if (!bfd_elf64_write_shdrs_and_ehdr (abfd))
return FALSE;
bfd_putl64 (elf_ia64_vms_tdata (abfd)->needed_count, needed_count);
if (bfd_seek (abfd, sizeof (Elf64_External_Ehdr), SEEK_SET) != 0
|| bfd_bwrite (needed_count, 8, abfd) != 8)
return FALSE;
return TRUE;
}
static bfd_boolean
elf64_vms_close_and_cleanup (bfd *abfd)
{
if (bfd_get_format (abfd) == bfd_object)
{
long isize;
/* Pad to 8 byte boundary for IPF/VMS. */
isize = bfd_get_size (abfd);
if ((isize & 7) != 0)
{
int ishort = 8 - (isize & 7);
bfd_uint64_t pad = 0;
bfd_seek (abfd, isize, SEEK_SET);
bfd_bwrite (&pad, ishort, abfd);
}
}
return _bfd_elf_close_and_cleanup (abfd);
}
/* Add symbols from an ELF object file to the linker hash table. */
static bfd_boolean
elf64_vms_link_add_object_symbols (bfd *abfd, struct bfd_link_info *info)
{
Elf_Internal_Shdr *hdr;
bfd_size_type symcount;
bfd_size_type extsymcount;
bfd_size_type extsymoff;
struct elf_link_hash_entry **sym_hash;
bfd_boolean dynamic;
Elf_Internal_Sym *isymbuf = NULL;
Elf_Internal_Sym *isym;
Elf_Internal_Sym *isymend;
const struct elf_backend_data *bed;
struct elf_link_hash_table *htab;
bfd_size_type amt;
htab = elf_hash_table (info);
bed = get_elf_backend_data (abfd);
if ((abfd->flags & DYNAMIC) == 0)
dynamic = FALSE;
else
{
dynamic = TRUE;
/* You can't use -r against a dynamic object. Also, there's no
hope of using a dynamic object which does not exactly match
the format of the output file. */
if (info->relocatable
|| !is_elf_hash_table (htab)
|| info->output_bfd->xvec != abfd->xvec)
{
if (info->relocatable)
bfd_set_error (bfd_error_invalid_operation);
else
bfd_set_error (bfd_error_wrong_format);
goto error_return;
}
}
if (! dynamic)
{
/* If we are creating a shared library, create all the dynamic
sections immediately. We need to attach them to something,
so we attach them to this BFD, provided it is the right
format. FIXME: If there are no input BFD's of the same
format as the output, we can't make a shared library. */
if (info->shared
&& is_elf_hash_table (htab)
&& info->output_bfd->xvec == abfd->xvec
&& !htab->dynamic_sections_created)
{
if (! elf64_ia64_create_dynamic_sections (abfd, info))
goto error_return;
}
}
else if (!is_elf_hash_table (htab))
goto error_return;
else
{
asection *s;
bfd_byte *dynbuf;
bfd_byte *extdyn;
/* ld --just-symbols and dynamic objects don't mix very well.
ld shouldn't allow it. */
if ((s = abfd->sections) != NULL
&& s->sec_info_type == SEC_INFO_TYPE_JUST_SYMS)
abort ();
/* Be sure there are dynamic sections. */
if (! elf64_ia64_create_dynamic_sections (htab->dynobj, info))
goto error_return;
s = bfd_get_section_by_name (abfd, ".dynamic");
if (s == NULL)
{
/* VMS libraries do not have dynamic sections. Create one from
the segment. */
Elf_Internal_Phdr *phdr;
unsigned int i, phnum;
phdr = elf_tdata (abfd)->phdr;
if (phdr == NULL)
goto error_return;
phnum = elf_elfheader (abfd)->e_phnum;
for (i = 0; i < phnum; phdr++)
if (phdr->p_type == PT_DYNAMIC)
{
s = bfd_make_section (abfd, ".dynamic");
if (s == NULL)
goto error_return;
s->vma = phdr->p_vaddr;
s->lma = phdr->p_paddr;
s->size = phdr->p_filesz;
s->filepos = phdr->p_offset;
s->flags |= SEC_HAS_CONTENTS;
s->alignment_power = bfd_log2 (phdr->p_align);
break;
}
if (s == NULL)
goto error_return;
}
/* Extract IDENT. */
if (!bfd_malloc_and_get_section (abfd, s, &dynbuf))
{
error_free_dyn:
free (dynbuf);
goto error_return;
}
for (extdyn = dynbuf;
extdyn < dynbuf + s->size;
extdyn += bed->s->sizeof_dyn)
{
Elf_Internal_Dyn dyn;
bed->s->swap_dyn_in (abfd, extdyn, &dyn);
if (dyn.d_tag == DT_IA_64_VMS_IDENT)
{
bfd_uint64_t tagv = dyn.d_un.d_val;
elf_ia64_vms_ident (abfd) = tagv;
break;
}
}
if (extdyn >= dynbuf + s->size)
{
/* Ident not found. */
goto error_free_dyn;
}
free (dynbuf);
/* We do not want to include any of the sections in a dynamic
object in the output file. We hack by simply clobbering the
list of sections in the BFD. This could be handled more
cleanly by, say, a new section flag; the existing
SEC_NEVER_LOAD flag is not the one we want, because that one
still implies that the section takes up space in the output
file. */
bfd_section_list_clear (abfd);
/* FIXME: should we detect if this library is already included ?
This should be harmless and shouldn't happen in practice. */
}
hdr = &elf_tdata (abfd)->symtab_hdr;
symcount = hdr->sh_size / bed->s->sizeof_sym;
/* The sh_info field of the symtab header tells us where the
external symbols start. We don't care about the local symbols at
this point. */
extsymcount = symcount - hdr->sh_info;
extsymoff = hdr->sh_info;
sym_hash = NULL;
if (extsymcount != 0)
{
isymbuf = bfd_elf_get_elf_syms (abfd, hdr, extsymcount, extsymoff,
NULL, NULL, NULL);
if (isymbuf == NULL)
goto error_return;
/* We store a pointer to the hash table entry for each external
symbol. */
amt = extsymcount * sizeof (struct elf_link_hash_entry *);
sym_hash = (struct elf_link_hash_entry **) bfd_alloc (abfd, amt);
if (sym_hash == NULL)
goto error_free_sym;
elf_sym_hashes (abfd) = sym_hash;
}
for (isym = isymbuf, isymend = isymbuf + extsymcount;
isym < isymend;
isym++, sym_hash++)
{
int bind;
bfd_vma value;
asection *sec, *new_sec;
flagword flags;
const char *name;
struct elf_link_hash_entry *h;
bfd_boolean definition;
bfd_boolean size_change_ok;
bfd_boolean type_change_ok;
bfd_boolean common;
unsigned int old_alignment;
bfd *old_bfd;
flags = BSF_NO_FLAGS;
sec = NULL;
value = isym->st_value;
*sym_hash = NULL;
common = bed->common_definition (isym);
bind = ELF_ST_BIND (isym->st_info);
switch (bind)
{
case STB_LOCAL:
/* This should be impossible, since ELF requires that all
global symbols follow all local symbols, and that sh_info
point to the first global symbol. Unfortunately, Irix 5
screws this up. */
continue;
case STB_GLOBAL:
if (isym->st_shndx != SHN_UNDEF && !common)
flags = BSF_GLOBAL;
break;
case STB_WEAK:
flags = BSF_WEAK;
break;
case STB_GNU_UNIQUE:
flags = BSF_GNU_UNIQUE;
break;
default:
/* Leave it up to the processor backend. */
break;
}
if (isym->st_shndx == SHN_UNDEF)
sec = bfd_und_section_ptr;
else if (isym->st_shndx == SHN_ABS)
sec = bfd_abs_section_ptr;
else if (isym->st_shndx == SHN_COMMON)
{
sec = bfd_com_section_ptr;
/* What ELF calls the size we call the value. What ELF
calls the value we call the alignment. */
value = isym->st_size;
}
else
{
sec = bfd_section_from_elf_index (abfd, isym->st_shndx);
if (sec == NULL)
sec = bfd_abs_section_ptr;
else if (sec->kept_section)
{
/* Symbols from discarded section are undefined. We keep
its visibility. */
sec = bfd_und_section_ptr;
isym->st_shndx = SHN_UNDEF;
}
else if ((abfd->flags & (EXEC_P | DYNAMIC)) != 0)
value -= sec->vma;
}
name = bfd_elf_string_from_elf_section (abfd, hdr->sh_link,
isym->st_name);
if (name == NULL)
goto error_free_vers;
if (bed->elf_add_symbol_hook)
{
if (! (*bed->elf_add_symbol_hook) (abfd, info, isym, &name, &flags,
&sec, &value))
goto error_free_vers;
/* The hook function sets the name to NULL if this symbol
should be skipped for some reason. */
if (name == NULL)
continue;
}
/* Sanity check that all possibilities were handled. */
if (sec == NULL)
{
bfd_set_error (bfd_error_bad_value);
goto error_free_vers;
}
if (bfd_is_und_section (sec)
|| bfd_is_com_section (sec))
definition = FALSE;
else
definition = TRUE;
size_change_ok = FALSE;
type_change_ok = bed->type_change_ok;
old_alignment = 0;
old_bfd = NULL;
new_sec = sec;
if (! bfd_is_und_section (sec))
h = elf_link_hash_lookup (htab, name, TRUE, FALSE, FALSE);
else
h = ((struct elf_link_hash_entry *) bfd_wrapped_link_hash_lookup
(abfd, info, name, TRUE, FALSE, FALSE));
if (h == NULL)
goto error_free_sym;
*sym_hash = h;
if (is_elf_hash_table (htab))
{
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
/* Remember the old alignment if this is a common symbol, so
that we don't reduce the alignment later on. We can't
check later, because _bfd_generic_link_add_one_symbol
will set a default for the alignment which we want to
override. We also remember the old bfd where the existing
definition comes from. */
switch (h->root.type)
{
default:
break;
case bfd_link_hash_defined:
if (abfd->selective_search)
continue;
/* Fall-through. */
case bfd_link_hash_defweak:
old_bfd = h->root.u.def.section->owner;
break;
case bfd_link_hash_common:
old_bfd = h->root.u.c.p->section->owner;
old_alignment = h->root.u.c.p->alignment_power;
break;
}
}
if (! (_bfd_generic_link_add_one_symbol
(info, abfd, name, flags, sec, value, NULL, FALSE, bed->collect,
(struct bfd_link_hash_entry **) sym_hash)))
goto error_free_vers;
h = *sym_hash;
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
*sym_hash = h;
if (definition)
h->unique_global = (flags & BSF_GNU_UNIQUE) != 0;
/* Set the alignment of a common symbol. */
if ((common || bfd_is_com_section (sec))
&& h->root.type == bfd_link_hash_common)
{
unsigned int align;
if (common)
align = bfd_log2 (isym->st_value);
else
{
/* The new symbol is a common symbol in a shared object.
We need to get the alignment from the section. */
align = new_sec->alignment_power;
}
if (align > old_alignment
/* Permit an alignment power of zero if an alignment of one
is specified and no other alignments have been specified. */
|| (isym->st_value == 1 && old_alignment == 0))
h->root.u.c.p->alignment_power = align;
else
h->root.u.c.p->alignment_power = old_alignment;
}
if (is_elf_hash_table (htab))
{
/* Check the alignment when a common symbol is involved. This
can change when a common symbol is overridden by a normal
definition or a common symbol is ignored due to the old
normal definition. We need to make sure the maximum
alignment is maintained. */
if ((old_alignment || common)
&& h->root.type != bfd_link_hash_common)
{
unsigned int common_align;
unsigned int normal_align;
unsigned int symbol_align;
bfd *normal_bfd;
bfd *common_bfd;
symbol_align = ffs (h->root.u.def.value) - 1;
if (h->root.u.def.section->owner != NULL
&& (h->root.u.def.section->owner->flags & DYNAMIC) == 0)
{
normal_align = h->root.u.def.section->alignment_power;
if (normal_align > symbol_align)
normal_align = symbol_align;
}
else
normal_align = symbol_align;
if (old_alignment)
{
common_align = old_alignment;
common_bfd = old_bfd;
normal_bfd = abfd;
}
else
{
common_align = bfd_log2 (isym->st_value);
common_bfd = abfd;
normal_bfd = old_bfd;
}
if (normal_align < common_align)
{
/* PR binutils/2735 */
if (normal_bfd == NULL)
(*_bfd_error_handler)
(_("Warning: alignment %u of common symbol `%s' in %B"
" is greater than the alignment (%u) of its section %A"),
common_bfd, h->root.u.def.section,
1 << common_align, name, 1 << normal_align);
else
(*_bfd_error_handler)
(_("Warning: alignment %u of symbol `%s' in %B"
" is smaller than %u in %B"),
normal_bfd, common_bfd,
1 << normal_align, name, 1 << common_align);
}
}
/* Remember the symbol size if it isn't undefined. */
if ((isym->st_size != 0 && isym->st_shndx != SHN_UNDEF)
&& (definition || h->size == 0))
{
if (h->size != 0
&& h->size != isym->st_size
&& ! size_change_ok)
(*_bfd_error_handler)
(_("Warning: size of symbol `%s' changed"
" from %lu in %B to %lu in %B"),
old_bfd, abfd,
name, (unsigned long) h->size,
(unsigned long) isym->st_size);
h->size = isym->st_size;
}
/* If this is a common symbol, then we always want H->SIZE
to be the size of the common symbol. The code just above
won't fix the size if a common symbol becomes larger. We
don't warn about a size change here, because that is
covered by --warn-common. Allow changed between different
function types. */
if (h->root.type == bfd_link_hash_common)
h->size = h->root.u.c.size;
if (ELF_ST_TYPE (isym->st_info) != STT_NOTYPE
&& (definition || h->type == STT_NOTYPE))
{
unsigned int type = ELF_ST_TYPE (isym->st_info);
if (h->type != type)
{
if (h->type != STT_NOTYPE && ! type_change_ok)
(*_bfd_error_handler)
(_("Warning: type of symbol `%s' changed"
" from %d to %d in %B"),
abfd, name, h->type, type);
h->type = type;
}
}
/* Set a flag in the hash table entry indicating the type of
reference or definition we just found. Keep a count of
the number of dynamic symbols we find. A dynamic symbol
is one which is referenced or defined by both a regular
object and a shared object. */
if (! dynamic)
{
if (! definition)
{
h->ref_regular = 1;
if (bind != STB_WEAK)
h->ref_regular_nonweak = 1;
}
else
{
BFD_ASSERT (!h->def_dynamic);
h->def_regular = 1;
}
}
else
{
BFD_ASSERT (definition);
h->def_dynamic = 1;
h->dynindx = -2;
((struct elf64_ia64_link_hash_entry *)h)->shl = abfd;
}
}
}
if (isymbuf != NULL)
{
free (isymbuf);
isymbuf = NULL;
}
/* If this object is the same format as the output object, and it is
not a shared library, then let the backend look through the
relocs.
This is required to build global offset table entries and to
arrange for dynamic relocs. It is not required for the
particular common case of linking non PIC code, even when linking
against shared libraries, but unfortunately there is no way of
knowing whether an object file has been compiled PIC or not.
Looking through the relocs is not particularly time consuming.
The problem is that we must either (1) keep the relocs in memory,
which causes the linker to require additional runtime memory or
(2) read the relocs twice from the input file, which wastes time.
This would be a good case for using mmap.
I have no idea how to handle linking PIC code into a file of a
different format. It probably can't be done. */
if (! dynamic
&& is_elf_hash_table (htab)
&& bed->check_relocs != NULL
&& (*bed->relocs_compatible) (abfd->xvec, info->output_bfd->xvec))
{
asection *o;
for (o = abfd->sections; o != NULL; o = o->next)
{
Elf_Internal_Rela *internal_relocs;
bfd_boolean ok;
if ((o->flags & SEC_RELOC) == 0
|| o->reloc_count == 0
|| ((info->strip == strip_all || info->strip == strip_debugger)
&& (o->flags & SEC_DEBUGGING) != 0)
|| bfd_is_abs_section (o->output_section))
continue;
internal_relocs = _bfd_elf_link_read_relocs (abfd, o, NULL, NULL,
info->keep_memory);
if (internal_relocs == NULL)
goto error_return;
ok = (*bed->check_relocs) (abfd, info, o, internal_relocs);
if (elf_section_data (o)->relocs != internal_relocs)
free (internal_relocs);
if (! ok)
goto error_return;
}
}
return TRUE;
error_free_vers:
error_free_sym:
if (isymbuf != NULL)
free (isymbuf);
error_return:
return FALSE;
}
static bfd_boolean
elf64_vms_link_add_archive_symbols (bfd *abfd, struct bfd_link_info *info)
{
int pass;
struct bfd_link_hash_entry **pundef;
struct bfd_link_hash_entry **next_pundef;
/* We only accept VMS libraries. */
if (info->output_bfd->xvec != abfd->xvec)
{
bfd_set_error (bfd_error_wrong_format);
return FALSE;
}
/* The archive_pass field in the archive itself is used to
initialize PASS, since we may search the same archive multiple
times. */
pass = ++abfd->archive_pass;
/* Look through the list of undefined symbols. */
for (pundef = &info->hash->undefs; *pundef != NULL; pundef = next_pundef)
{
struct bfd_link_hash_entry *h;
symindex symidx;
bfd *element;
bfd *orig_element;
h = *pundef;
next_pundef = &(*pundef)->u.undef.next;
/* When a symbol is defined, it is not necessarily removed from
the list. */
if (h->type != bfd_link_hash_undefined
&& h->type != bfd_link_hash_common)
{
/* Remove this entry from the list, for general cleanliness
and because we are going to look through the list again
if we search any more libraries. We can't remove the
entry if it is the tail, because that would lose any
entries we add to the list later on. */
if (*pundef != info->hash->undefs_tail)
{
*pundef = *next_pundef;
next_pundef = pundef;
}
continue;
}
/* Look for this symbol in the archive hash table. */
symidx = _bfd_vms_lib_find_symbol (abfd, h->root.string);
if (symidx == BFD_NO_MORE_SYMBOLS)
{
/* Nothing in this slot. */
continue;
}
element = bfd_get_elt_at_index (abfd, symidx);
if (element == NULL)
return FALSE;
if (element->archive_pass == -1 || element->archive_pass == pass)
{
/* Next symbol if this archive is wrong or already handled. */
continue;
}
orig_element = element;
if (bfd_is_thin_archive (abfd))
{
element = _bfd_vms_lib_get_imagelib_file (element);
if (element == NULL || !bfd_check_format (element, bfd_object))
{
orig_element->archive_pass = -1;
return FALSE;
}
}
else if (! bfd_check_format (element, bfd_object))
{
element->archive_pass = -1;
return FALSE;
}
/* Unlike the generic linker, we know that this element provides
a definition for an undefined symbol and we know that we want
to include it. We don't need to check anything. */
if (! (*info->callbacks->add_archive_element) (info, element,
h->root.string, &element))
return FALSE;
if (! elf64_vms_link_add_object_symbols (element, info))
return FALSE;
orig_element->archive_pass = pass;
}
return TRUE;
}
static bfd_boolean
elf64_vms_bfd_link_add_symbols (bfd *abfd, struct bfd_link_info *info)
{
switch (bfd_get_format (abfd))
{
case bfd_object:
return elf64_vms_link_add_object_symbols (abfd, info);
break;
case bfd_archive:
return elf64_vms_link_add_archive_symbols (abfd, info);
break;
default:
bfd_set_error (bfd_error_wrong_format);
return FALSE;
}
}
static bfd_boolean
elf64_ia64_vms_mkobject (bfd *abfd)
{
return bfd_elf_allocate_object
(abfd, sizeof (struct elf64_ia64_vms_obj_tdata), IA64_ELF_DATA);
}
/* Size-dependent data and functions. */
static const struct elf_size_info elf64_ia64_vms_size_info = {
sizeof (Elf64_External_VMS_Ehdr),
sizeof (Elf64_External_Phdr),
sizeof (Elf64_External_Shdr),
sizeof (Elf64_External_Rel),
sizeof (Elf64_External_Rela),
sizeof (Elf64_External_Sym),
sizeof (Elf64_External_Dyn),
sizeof (Elf_External_Note),
4,
1,
64, 3, /* ARCH_SIZE, LOG_FILE_ALIGN */
ELFCLASS64, EV_CURRENT,
bfd_elf64_write_out_phdrs,
elf64_vms_write_shdrs_and_ehdr,
bfd_elf64_checksum_contents,
bfd_elf64_write_relocs,
bfd_elf64_swap_symbol_in,
bfd_elf64_swap_symbol_out,
bfd_elf64_slurp_reloc_table,
bfd_elf64_slurp_symbol_table,
bfd_elf64_swap_dyn_in,
bfd_elf64_swap_dyn_out,
bfd_elf64_swap_reloc_in,
bfd_elf64_swap_reloc_out,
bfd_elf64_swap_reloca_in,
bfd_elf64_swap_reloca_out
};
#define ELF_ARCH bfd_arch_ia64
#define ELF_MACHINE_CODE EM_IA_64
#define ELF_MAXPAGESIZE 0x10000 /* 64KB */
#define ELF_COMMONPAGESIZE 0x200 /* 16KB */
#define elf_backend_section_from_shdr \
elf64_ia64_section_from_shdr
#define elf_backend_section_flags \
elf64_ia64_section_flags
#define elf_backend_fake_sections \
elf64_ia64_fake_sections
#define elf_backend_final_write_processing \
elf64_ia64_final_write_processing
#define elf_backend_add_symbol_hook \
elf64_ia64_add_symbol_hook
#define elf_info_to_howto \
elf64_ia64_info_to_howto
#define bfd_elf64_bfd_reloc_type_lookup \
ia64_elf_reloc_type_lookup
#define bfd_elf64_bfd_reloc_name_lookup \
ia64_elf_reloc_name_lookup
#define bfd_elf64_bfd_is_local_label_name \
elf64_ia64_is_local_label_name
#define bfd_elf64_bfd_relax_section \
elf64_ia64_relax_section
#define elf_backend_object_p \
elf64_ia64_object_p
/* Stuff for the BFD linker: */
#define bfd_elf64_bfd_link_hash_table_create \
elf64_ia64_hash_table_create
#define bfd_elf64_bfd_link_hash_table_free \
elf64_ia64_hash_table_free
#define elf_backend_create_dynamic_sections \
elf64_ia64_create_dynamic_sections
#define elf_backend_check_relocs \
elf64_ia64_check_relocs
#define elf_backend_adjust_dynamic_symbol \
elf64_ia64_adjust_dynamic_symbol
#define elf_backend_size_dynamic_sections \
elf64_ia64_size_dynamic_sections
#define elf_backend_omit_section_dynsym \
((bfd_boolean (*) (bfd *, struct bfd_link_info *, asection *)) bfd_true)
#define elf_backend_relocate_section \
elf64_ia64_relocate_section
#define elf_backend_finish_dynamic_symbol \
elf64_ia64_finish_dynamic_symbol
#define elf_backend_finish_dynamic_sections \
elf64_ia64_finish_dynamic_sections
#define bfd_elf64_bfd_final_link \
elf64_ia64_final_link
#define bfd_elf64_bfd_merge_private_bfd_data \
elf64_ia64_merge_private_bfd_data
#define bfd_elf64_bfd_set_private_flags \
elf64_ia64_set_private_flags
#define bfd_elf64_bfd_print_private_bfd_data \
elf64_ia64_print_private_bfd_data
#define elf_backend_plt_readonly 1
#define elf_backend_want_plt_sym 0
#define elf_backend_plt_alignment 5
#define elf_backend_got_header_size 0
#define elf_backend_want_got_plt 1
#define elf_backend_may_use_rel_p 1
#define elf_backend_may_use_rela_p 1
#define elf_backend_default_use_rela_p 1
#define elf_backend_want_dynbss 0
#define elf_backend_hide_symbol elf64_ia64_hash_hide_symbol
#define elf_backend_fixup_symbol _bfd_elf_link_hash_fixup_symbol
#define elf_backend_reloc_type_class elf64_ia64_reloc_type_class
#define elf_backend_rela_normal 1
#define elf_backend_special_sections elf64_ia64_special_sections
#define elf_backend_default_execstack 0
/* FIXME: PR 290: The Intel C compiler generates SHT_IA_64_UNWIND with
SHF_LINK_ORDER. But it doesn't set the sh_link or sh_info fields.
We don't want to flood users with so many error messages. We turn
off the warning for now. It will be turned on later when the Intel
compiler is fixed. */
#define elf_backend_link_order_error_handler NULL
/* VMS-specific vectors. */
#undef TARGET_LITTLE_SYM
#define TARGET_LITTLE_SYM bfd_elf64_ia64_vms_vec
#undef TARGET_LITTLE_NAME
#define TARGET_LITTLE_NAME "elf64-ia64-vms"
#undef TARGET_BIG_SYM
#undef TARGET_BIG_NAME
/* These are VMS specific functions. */
#undef elf_backend_object_p
#define elf_backend_object_p elf64_vms_object_p
#undef elf_backend_section_from_shdr
#define elf_backend_section_from_shdr elf64_vms_section_from_shdr
#undef elf_backend_post_process_headers
#define elf_backend_post_process_headers elf64_vms_post_process_headers
#undef elf_backend_section_processing
#define elf_backend_section_processing elf64_vms_section_processing
#undef elf_backend_final_write_processing
#define elf_backend_final_write_processing elf64_vms_final_write_processing
#undef bfd_elf64_close_and_cleanup
#define bfd_elf64_close_and_cleanup elf64_vms_close_and_cleanup
#undef elf_backend_section_from_bfd_section
#undef elf_backend_symbol_processing
#undef elf_backend_want_p_paddr_set_to_zero
#undef ELF_OSABI
#define ELF_OSABI ELFOSABI_OPENVMS
#undef ELF_MAXPAGESIZE
#define ELF_MAXPAGESIZE 0x10000 /* 64KB */
#undef elf64_bed
#define elf64_bed elf64_ia64_vms_bed
#define elf_backend_size_info elf64_ia64_vms_size_info
/* Use VMS-style archives (in particular, don't use the standard coff
archive format). */
#define bfd_elf64_archive_functions
#undef bfd_elf64_archive_p
#define bfd_elf64_archive_p _bfd_vms_lib_ia64_archive_p
#undef bfd_elf64_write_archive_contents
#define bfd_elf64_write_archive_contents _bfd_vms_lib_write_archive_contents
#undef bfd_elf64_mkarchive
#define bfd_elf64_mkarchive _bfd_vms_lib_ia64_mkarchive
#define bfd_elf64_archive_slurp_armap \
_bfd_vms_lib_slurp_armap
#define bfd_elf64_archive_slurp_extended_name_table \
_bfd_vms_lib_slurp_extended_name_table
#define bfd_elf64_archive_construct_extended_name_table \
_bfd_vms_lib_construct_extended_name_table
#define bfd_elf64_archive_truncate_arname \
_bfd_vms_lib_truncate_arname
#define bfd_elf64_archive_write_armap \
_bfd_vms_lib_write_armap
#define bfd_elf64_archive_read_ar_hdr \
_bfd_vms_lib_read_ar_hdr
#define bfd_elf64_archive_write_ar_hdr \
_bfd_vms_lib_write_ar_hdr
#define bfd_elf64_archive_openr_next_archived_file \
_bfd_vms_lib_openr_next_archived_file
#define bfd_elf64_archive_get_elt_at_index \
_bfd_vms_lib_get_elt_at_index
#define bfd_elf64_archive_generic_stat_arch_elt \
_bfd_vms_lib_generic_stat_arch_elt
#define bfd_elf64_archive_update_armap_timestamp \
_bfd_vms_lib_update_armap_timestamp
/* VMS link methods. */
#undef bfd_elf64_bfd_link_add_symbols
#define bfd_elf64_bfd_link_add_symbols elf64_vms_bfd_link_add_symbols
#undef elf_backend_want_got_sym
#define elf_backend_want_got_sym 0
#undef bfd_elf64_mkobject
#define bfd_elf64_mkobject elf64_ia64_vms_mkobject
/* Redefine to align segments on block size. */
#undef ELF_MAXPAGESIZE
#define ELF_MAXPAGESIZE 0x200 /* 512B */
#undef elf_backend_want_got_plt
#define elf_backend_want_got_plt 0
#include "elf64-target.h"