// arm.cc -- arm target support for gold. // Copyright 2009, 2010 Free Software Foundation, Inc. // Written by Doug Kwan based on the i386 code // by Ian Lance Taylor . // This file also contains borrowed and adapted code from // bfd/elf32-arm.c. // This file is part of gold. // This program is free software; you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation; either version 3 of the License, or // (at your option) any later version. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program; if not, write to the Free Software // Foundation, Inc., 51 Franklin Street - Fifth Floor, Boston, // MA 02110-1301, USA. #include "gold.h" #include #include #include #include #include #include #include #include #include "elfcpp.h" #include "parameters.h" #include "reloc.h" #include "arm.h" #include "object.h" #include "symtab.h" #include "layout.h" #include "output.h" #include "copy-relocs.h" #include "target.h" #include "target-reloc.h" #include "target-select.h" #include "tls.h" #include "defstd.h" #include "gc.h" #include "attributes.h" #include "arm-reloc-property.h" namespace { using namespace gold; template class Output_data_plt_arm; template class Stub_table; template class Arm_input_section; class Arm_exidx_cantunwind; class Arm_exidx_merged_section; class Arm_exidx_fixup; template class Arm_output_section; class Arm_exidx_input_section; template class Arm_relobj; template class Target_arm; // For convenience. typedef elfcpp::Elf_types<32>::Elf_Addr Arm_address; // Maximum branch offsets for ARM, THUMB and THUMB2. const int32_t ARM_MAX_FWD_BRANCH_OFFSET = ((((1 << 23) - 1) << 2) + 8); const int32_t ARM_MAX_BWD_BRANCH_OFFSET = ((-((1 << 23) << 2)) + 8); const int32_t THM_MAX_FWD_BRANCH_OFFSET = ((1 << 22) -2 + 4); const int32_t THM_MAX_BWD_BRANCH_OFFSET = (-(1 << 22) + 4); const int32_t THM2_MAX_FWD_BRANCH_OFFSET = (((1 << 24) - 2) + 4); const int32_t THM2_MAX_BWD_BRANCH_OFFSET = (-(1 << 24) + 4); // The arm target class. // // This is a very simple port of gold for ARM-EABI. It is intended for // supporting Android only for the time being. // // TODOs: // - Implement all static relocation types documented in arm-reloc.def. // - Make PLTs more flexible for different architecture features like // Thumb-2 and BE8. // There are probably a lot more. // Ideally we would like to avoid using global variables but this is used // very in many places and sometimes in loops. If we use a function // returning a static instance of Arm_reloc_property_table, it will very // slow in an threaded environment since the static instance needs to be // locked. The pointer is below initialized in the // Target::do_select_as_default_target() hook so that we do not spend time // building the table if we are not linking ARM objects. // // An alternative is to to process the information in arm-reloc.def in // compilation time and generate a representation of it in PODs only. That // way we can avoid initialization when the linker starts. Arm_reloc_property_table *arm_reloc_property_table = NULL; // Instruction template class. This class is similar to the insn_sequence // struct in bfd/elf32-arm.c. class Insn_template { public: // Types of instruction templates. enum Type { THUMB16_TYPE = 1, // THUMB16_SPECIAL_TYPE is used by sub-classes of Stub for instruction // templates with class-specific semantics. Currently this is used // only by the Cortex_a8_stub class for handling condition codes in // conditional branches. THUMB16_SPECIAL_TYPE, THUMB32_TYPE, ARM_TYPE, DATA_TYPE }; // Factory methods to create instruction templates in different formats. static const Insn_template thumb16_insn(uint32_t data) { return Insn_template(data, THUMB16_TYPE, elfcpp::R_ARM_NONE, 0); } // A Thumb conditional branch, in which the proper condition is inserted // when we build the stub. static const Insn_template thumb16_bcond_insn(uint32_t data) { return Insn_template(data, THUMB16_SPECIAL_TYPE, elfcpp::R_ARM_NONE, 1); } static const Insn_template thumb32_insn(uint32_t data) { return Insn_template(data, THUMB32_TYPE, elfcpp::R_ARM_NONE, 0); } static const Insn_template thumb32_b_insn(uint32_t data, int reloc_addend) { return Insn_template(data, THUMB32_TYPE, elfcpp::R_ARM_THM_JUMP24, reloc_addend); } static const Insn_template arm_insn(uint32_t data) { return Insn_template(data, ARM_TYPE, elfcpp::R_ARM_NONE, 0); } static const Insn_template arm_rel_insn(unsigned data, int reloc_addend) { return Insn_template(data, ARM_TYPE, elfcpp::R_ARM_JUMP24, reloc_addend); } static const Insn_template data_word(unsigned data, unsigned int r_type, int reloc_addend) { return Insn_template(data, DATA_TYPE, r_type, reloc_addend); } // Accessors. This class is used for read-only objects so no modifiers // are provided. uint32_t data() const { return this->data_; } // Return the instruction sequence type of this. Type type() const { return this->type_; } // Return the ARM relocation type of this. unsigned int r_type() const { return this->r_type_; } int32_t reloc_addend() const { return this->reloc_addend_; } // Return size of instruction template in bytes. size_t size() const; // Return byte-alignment of instruction template. unsigned alignment() const; private: // We make the constructor private to ensure that only the factory // methods are used. inline Insn_template(unsigned data, Type type, unsigned int r_type, int reloc_addend) : data_(data), type_(type), r_type_(r_type), reloc_addend_(reloc_addend) { } // Instruction specific data. This is used to store information like // some of the instruction bits. uint32_t data_; // Instruction template type. Type type_; // Relocation type if there is a relocation or R_ARM_NONE otherwise. unsigned int r_type_; // Relocation addend. int32_t reloc_addend_; }; // Macro for generating code to stub types. One entry per long/short // branch stub #define DEF_STUBS \ DEF_STUB(long_branch_any_any) \ DEF_STUB(long_branch_v4t_arm_thumb) \ DEF_STUB(long_branch_thumb_only) \ DEF_STUB(long_branch_v4t_thumb_thumb) \ DEF_STUB(long_branch_v4t_thumb_arm) \ DEF_STUB(short_branch_v4t_thumb_arm) \ DEF_STUB(long_branch_any_arm_pic) \ DEF_STUB(long_branch_any_thumb_pic) \ DEF_STUB(long_branch_v4t_thumb_thumb_pic) \ DEF_STUB(long_branch_v4t_arm_thumb_pic) \ DEF_STUB(long_branch_v4t_thumb_arm_pic) \ DEF_STUB(long_branch_thumb_only_pic) \ DEF_STUB(a8_veneer_b_cond) \ DEF_STUB(a8_veneer_b) \ DEF_STUB(a8_veneer_bl) \ DEF_STUB(a8_veneer_blx) \ DEF_STUB(v4_veneer_bx) // Stub types. #define DEF_STUB(x) arm_stub_##x, typedef enum { arm_stub_none, DEF_STUBS // First reloc stub type. arm_stub_reloc_first = arm_stub_long_branch_any_any, // Last reloc stub type. arm_stub_reloc_last = arm_stub_long_branch_thumb_only_pic, // First Cortex-A8 stub type. arm_stub_cortex_a8_first = arm_stub_a8_veneer_b_cond, // Last Cortex-A8 stub type. arm_stub_cortex_a8_last = arm_stub_a8_veneer_blx, // Last stub type. arm_stub_type_last = arm_stub_v4_veneer_bx } Stub_type; #undef DEF_STUB // Stub template class. Templates are meant to be read-only objects. // A stub template for a stub type contains all read-only attributes // common to all stubs of the same type. class Stub_template { public: Stub_template(Stub_type, const Insn_template*, size_t); ~Stub_template() { } // Return stub type. Stub_type type() const { return this->type_; } // Return an array of instruction templates. const Insn_template* insns() const { return this->insns_; } // Return size of template in number of instructions. size_t insn_count() const { return this->insn_count_; } // Return size of template in bytes. size_t size() const { return this->size_; } // Return alignment of the stub template. unsigned alignment() const { return this->alignment_; } // Return whether entry point is in thumb mode. bool entry_in_thumb_mode() const { return this->entry_in_thumb_mode_; } // Return number of relocations in this template. size_t reloc_count() const { return this->relocs_.size(); } // Return index of the I-th instruction with relocation. size_t reloc_insn_index(size_t i) const { gold_assert(i < this->relocs_.size()); return this->relocs_[i].first; } // Return the offset of the I-th instruction with relocation from the // beginning of the stub. section_size_type reloc_offset(size_t i) const { gold_assert(i < this->relocs_.size()); return this->relocs_[i].second; } private: // This contains information about an instruction template with a relocation // and its offset from start of stub. typedef std::pair Reloc; // A Stub_template may not be copied. We want to share templates as much // as possible. Stub_template(const Stub_template&); Stub_template& operator=(const Stub_template&); // Stub type. Stub_type type_; // Points to an array of Insn_templates. const Insn_template* insns_; // Number of Insn_templates in insns_[]. size_t insn_count_; // Size of templated instructions in bytes. size_t size_; // Alignment of templated instructions. unsigned alignment_; // Flag to indicate if entry is in thumb mode. bool entry_in_thumb_mode_; // A table of reloc instruction indices and offsets. We can find these by // looking at the instruction templates but we pre-compute and then stash // them here for speed. std::vector relocs_; }; // // A class for code stubs. This is a base class for different type of // stubs used in the ARM target. // class Stub { private: static const section_offset_type invalid_offset = static_cast(-1); public: Stub(const Stub_template* stub_template) : stub_template_(stub_template), offset_(invalid_offset) { } virtual ~Stub() { } // Return the stub template. const Stub_template* stub_template() const { return this->stub_template_; } // Return offset of code stub from beginning of its containing stub table. section_offset_type offset() const { gold_assert(this->offset_ != invalid_offset); return this->offset_; } // Set offset of code stub from beginning of its containing stub table. void set_offset(section_offset_type offset) { this->offset_ = offset; } // Return the relocation target address of the i-th relocation in the // stub. This must be defined in a child class. Arm_address reloc_target(size_t i) { return this->do_reloc_target(i); } // Write a stub at output VIEW. BIG_ENDIAN select how a stub is written. void write(unsigned char* view, section_size_type view_size, bool big_endian) { this->do_write(view, view_size, big_endian); } // Return the instruction for THUMB16_SPECIAL_TYPE instruction template // for the i-th instruction. uint16_t thumb16_special(size_t i) { return this->do_thumb16_special(i); } protected: // This must be defined in the child class. virtual Arm_address do_reloc_target(size_t) = 0; // This may be overridden in the child class. virtual void do_write(unsigned char* view, section_size_type view_size, bool big_endian) { if (big_endian) this->do_fixed_endian_write(view, view_size); else this->do_fixed_endian_write(view, view_size); } // This must be overridden if a child class uses the THUMB16_SPECIAL_TYPE // instruction template. virtual uint16_t do_thumb16_special(size_t) { gold_unreachable(); } private: // A template to implement do_write. template void inline do_fixed_endian_write(unsigned char*, section_size_type); // Its template. const Stub_template* stub_template_; // Offset within the section of containing this stub. section_offset_type offset_; }; // Reloc stub class. These are stubs we use to fix up relocation because // of limited branch ranges. class Reloc_stub : public Stub { public: static const unsigned int invalid_index = static_cast(-1); // We assume we never jump to this address. static const Arm_address invalid_address = static_cast(-1); // Return destination address. Arm_address destination_address() const { gold_assert(this->destination_address_ != this->invalid_address); return this->destination_address_; } // Set destination address. void set_destination_address(Arm_address address) { gold_assert(address != this->invalid_address); this->destination_address_ = address; } // Reset destination address. void reset_destination_address() { this->destination_address_ = this->invalid_address; } // Determine stub type for a branch of a relocation of R_TYPE going // from BRANCH_ADDRESS to BRANCH_TARGET. If TARGET_IS_THUMB is set, // the branch target is a thumb instruction. TARGET is used for look // up ARM-specific linker settings. static Stub_type stub_type_for_reloc(unsigned int r_type, Arm_address branch_address, Arm_address branch_target, bool target_is_thumb); // Reloc_stub key. A key is logically a triplet of a stub type, a symbol // and an addend. Since we treat global and local symbol differently, we // use a Symbol object for a global symbol and a object-index pair for // a local symbol. class Key { public: // If SYMBOL is not null, this is a global symbol, we ignore RELOBJ and // R_SYM. Otherwise, this is a local symbol and RELOBJ must non-NULL // and R_SYM must not be invalid_index. Key(Stub_type stub_type, const Symbol* symbol, const Relobj* relobj, unsigned int r_sym, int32_t addend) : stub_type_(stub_type), addend_(addend) { if (symbol != NULL) { this->r_sym_ = Reloc_stub::invalid_index; this->u_.symbol = symbol; } else { gold_assert(relobj != NULL && r_sym != invalid_index); this->r_sym_ = r_sym; this->u_.relobj = relobj; } } ~Key() { } // Accessors: Keys are meant to be read-only object so no modifiers are // provided. // Return stub type. Stub_type stub_type() const { return this->stub_type_; } // Return the local symbol index or invalid_index. unsigned int r_sym() const { return this->r_sym_; } // Return the symbol if there is one. const Symbol* symbol() const { return this->r_sym_ == invalid_index ? this->u_.symbol : NULL; } // Return the relobj if there is one. const Relobj* relobj() const { return this->r_sym_ != invalid_index ? this->u_.relobj : NULL; } // Whether this equals to another key k. bool eq(const Key& k) const { return ((this->stub_type_ == k.stub_type_) && (this->r_sym_ == k.r_sym_) && ((this->r_sym_ != Reloc_stub::invalid_index) ? (this->u_.relobj == k.u_.relobj) : (this->u_.symbol == k.u_.symbol)) && (this->addend_ == k.addend_)); } // Return a hash value. size_t hash_value() const { return (this->stub_type_ ^ this->r_sym_ ^ gold::string_hash( (this->r_sym_ != Reloc_stub::invalid_index) ? this->u_.relobj->name().c_str() : this->u_.symbol->name()) ^ this->addend_); } // Functors for STL associative containers. struct hash { size_t operator()(const Key& k) const { return k.hash_value(); } }; struct equal_to { bool operator()(const Key& k1, const Key& k2) const { return k1.eq(k2); } }; // Name of key. This is mainly for debugging. std::string name() const; private: // Stub type. Stub_type stub_type_; // If this is a local symbol, this is the index in the defining object. // Otherwise, it is invalid_index for a global symbol. unsigned int r_sym_; // If r_sym_ is invalid index. This points to a global symbol. // Otherwise, this points a relobj. We used the unsized and target // independent Symbol and Relobj classes instead of Sized_symbol<32> and // Arm_relobj. This is done to avoid making the stub class a template // as most of the stub machinery is endianity-neutral. However, it // may require a bit of casting done by users of this class. union { const Symbol* symbol; const Relobj* relobj; } u_; // Addend associated with a reloc. int32_t addend_; }; protected: // Reloc_stubs are created via a stub factory. So these are protected. Reloc_stub(const Stub_template* stub_template) : Stub(stub_template), destination_address_(invalid_address) { } ~Reloc_stub() { } friend class Stub_factory; // Return the relocation target address of the i-th relocation in the // stub. Arm_address do_reloc_target(size_t i) { // All reloc stub have only one relocation. gold_assert(i == 0); return this->destination_address_; } private: // Address of destination. Arm_address destination_address_; }; // Cortex-A8 stub class. We need a Cortex-A8 stub to redirect any 32-bit // THUMB branch that meets the following conditions: // // 1. The branch straddles across a page boundary. i.e. lower 12-bit of // branch address is 0xffe. // 2. The branch target address is in the same page as the first word of the // branch. // 3. The branch follows a 32-bit instruction which is not a branch. // // To do the fix up, we need to store the address of the branch instruction // and its target at least. We also need to store the original branch // instruction bits for the condition code in a conditional branch. The // condition code is used in a special instruction template. We also want // to identify input sections needing Cortex-A8 workaround quickly. We store // extra information about object and section index of the code section // containing a branch being fixed up. The information is used to mark // the code section when we finalize the Cortex-A8 stubs. // class Cortex_a8_stub : public Stub { public: ~Cortex_a8_stub() { } // Return the object of the code section containing the branch being fixed // up. Relobj* relobj() const { return this->relobj_; } // Return the section index of the code section containing the branch being // fixed up. unsigned int shndx() const { return this->shndx_; } // Return the source address of stub. This is the address of the original // branch instruction. LSB is 1 always set to indicate that it is a THUMB // instruction. Arm_address source_address() const { return this->source_address_; } // Return the destination address of the stub. This is the branch taken // address of the original branch instruction. LSB is 1 if it is a THUMB // instruction address. Arm_address destination_address() const { return this->destination_address_; } // Return the instruction being fixed up. uint32_t original_insn() const { return this->original_insn_; } protected: // Cortex_a8_stubs are created via a stub factory. So these are protected. Cortex_a8_stub(const Stub_template* stub_template, Relobj* relobj, unsigned int shndx, Arm_address source_address, Arm_address destination_address, uint32_t original_insn) : Stub(stub_template), relobj_(relobj), shndx_(shndx), source_address_(source_address | 1U), destination_address_(destination_address), original_insn_(original_insn) { } friend class Stub_factory; // Return the relocation target address of the i-th relocation in the // stub. Arm_address do_reloc_target(size_t i) { if (this->stub_template()->type() == arm_stub_a8_veneer_b_cond) { // The conditional branch veneer has two relocations. gold_assert(i < 2); return i == 0 ? this->source_address_ + 4 : this->destination_address_; } else { // All other Cortex-A8 stubs have only one relocation. gold_assert(i == 0); return this->destination_address_; } } // Return an instruction for the THUMB16_SPECIAL_TYPE instruction template. uint16_t do_thumb16_special(size_t); private: // Object of the code section containing the branch being fixed up. Relobj* relobj_; // Section index of the code section containing the branch begin fixed up. unsigned int shndx_; // Source address of original branch. Arm_address source_address_; // Destination address of the original branch. Arm_address destination_address_; // Original branch instruction. This is needed for copying the condition // code from a condition branch to its stub. uint32_t original_insn_; }; // ARMv4 BX Rx branch relocation stub class. class Arm_v4bx_stub : public Stub { public: ~Arm_v4bx_stub() { } // Return the associated register. uint32_t reg() const { return this->reg_; } protected: // Arm V4BX stubs are created via a stub factory. So these are protected. Arm_v4bx_stub(const Stub_template* stub_template, const uint32_t reg) : Stub(stub_template), reg_(reg) { } friend class Stub_factory; // Return the relocation target address of the i-th relocation in the // stub. Arm_address do_reloc_target(size_t) { gold_unreachable(); } // This may be overridden in the child class. virtual void do_write(unsigned char* view, section_size_type view_size, bool big_endian) { if (big_endian) this->do_fixed_endian_v4bx_write(view, view_size); else this->do_fixed_endian_v4bx_write(view, view_size); } private: // A template to implement do_write. template void inline do_fixed_endian_v4bx_write(unsigned char* view, section_size_type) { const Insn_template* insns = this->stub_template()->insns(); elfcpp::Swap<32, big_endian>::writeval(view, (insns[0].data() + (this->reg_ << 16))); view += insns[0].size(); elfcpp::Swap<32, big_endian>::writeval(view, (insns[1].data() + this->reg_)); view += insns[1].size(); elfcpp::Swap<32, big_endian>::writeval(view, (insns[2].data() + this->reg_)); } // A register index (r0-r14), which is associated with the stub. uint32_t reg_; }; // Stub factory class. class Stub_factory { public: // Return the unique instance of this class. static const Stub_factory& get_instance() { static Stub_factory singleton; return singleton; } // Make a relocation stub. Reloc_stub* make_reloc_stub(Stub_type stub_type) const { gold_assert(stub_type >= arm_stub_reloc_first && stub_type <= arm_stub_reloc_last); return new Reloc_stub(this->stub_templates_[stub_type]); } // Make a Cortex-A8 stub. Cortex_a8_stub* make_cortex_a8_stub(Stub_type stub_type, Relobj* relobj, unsigned int shndx, Arm_address source, Arm_address destination, uint32_t original_insn) const { gold_assert(stub_type >= arm_stub_cortex_a8_first && stub_type <= arm_stub_cortex_a8_last); return new Cortex_a8_stub(this->stub_templates_[stub_type], relobj, shndx, source, destination, original_insn); } // Make an ARM V4BX relocation stub. // This method creates a stub from the arm_stub_v4_veneer_bx template only. Arm_v4bx_stub* make_arm_v4bx_stub(uint32_t reg) const { gold_assert(reg < 0xf); return new Arm_v4bx_stub(this->stub_templates_[arm_stub_v4_veneer_bx], reg); } private: // Constructor and destructor are protected since we only return a single // instance created in Stub_factory::get_instance(). Stub_factory(); // A Stub_factory may not be copied since it is a singleton. Stub_factory(const Stub_factory&); Stub_factory& operator=(Stub_factory&); // Stub templates. These are initialized in the constructor. const Stub_template* stub_templates_[arm_stub_type_last+1]; }; // A class to hold stubs for the ARM target. template class Stub_table : public Output_data { public: Stub_table(Arm_input_section* owner) : Output_data(), owner_(owner), reloc_stubs_(), cortex_a8_stubs_(), arm_v4bx_stubs_(0xf), prev_data_size_(0), prev_addralign_(1) { } ~Stub_table() { } // Owner of this stub table. Arm_input_section* owner() const { return this->owner_; } // Whether this stub table is empty. bool empty() const { return (this->reloc_stubs_.empty() && this->cortex_a8_stubs_.empty() && this->arm_v4bx_stubs_.empty()); } // Return the current data size. off_t current_data_size() const { return this->current_data_size_for_child(); } // Add a STUB with using KEY. Caller is reponsible for avoid adding // if already a STUB with the same key has been added. void add_reloc_stub(Reloc_stub* stub, const Reloc_stub::Key& key) { const Stub_template* stub_template = stub->stub_template(); gold_assert(stub_template->type() == key.stub_type()); this->reloc_stubs_[key] = stub; } // Add a Cortex-A8 STUB that fixes up a THUMB branch at ADDRESS. // Caller is reponsible for avoid adding if already a STUB with the same // address has been added. void add_cortex_a8_stub(Arm_address address, Cortex_a8_stub* stub) { std::pair value(address, stub); this->cortex_a8_stubs_.insert(value); } // Add an ARM V4BX relocation stub. A register index will be retrieved // from the stub. void add_arm_v4bx_stub(Arm_v4bx_stub* stub) { gold_assert(stub != NULL && this->arm_v4bx_stubs_[stub->reg()] == NULL); this->arm_v4bx_stubs_[stub->reg()] = stub; } // Remove all Cortex-A8 stubs. void remove_all_cortex_a8_stubs(); // Look up a relocation stub using KEY. Return NULL if there is none. Reloc_stub* find_reloc_stub(const Reloc_stub::Key& key) const { typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.find(key); return (p != this->reloc_stubs_.end()) ? p->second : NULL; } // Look up an arm v4bx relocation stub using the register index. // Return NULL if there is none. Arm_v4bx_stub* find_arm_v4bx_stub(const uint32_t reg) const { gold_assert(reg < 0xf); return this->arm_v4bx_stubs_[reg]; } // Relocate stubs in this stub table. void relocate_stubs(const Relocate_info<32, big_endian>*, Target_arm*, Output_section*, unsigned char*, Arm_address, section_size_type); // Update data size and alignment at the end of a relaxation pass. Return // true if either data size or alignment is different from that of the // previous relaxation pass. bool update_data_size_and_addralign(); // Finalize stubs. Set the offsets of all stubs and mark input sections // needing the Cortex-A8 workaround. void finalize_stubs(); // Apply Cortex-A8 workaround to an address range. void apply_cortex_a8_workaround_to_address_range(Target_arm*, unsigned char*, Arm_address, section_size_type); protected: // Write out section contents. void do_write(Output_file*); // Return the required alignment. uint64_t do_addralign() const { return this->prev_addralign_; } // Reset address and file offset. void do_reset_address_and_file_offset() { this->set_current_data_size_for_child(this->prev_data_size_); } // Set final data size. void set_final_data_size() { this->set_data_size(this->current_data_size()); } private: // Relocate one stub. void relocate_stub(Stub*, const Relocate_info<32, big_endian>*, Target_arm*, Output_section*, unsigned char*, Arm_address, section_size_type); // Unordered map of relocation stubs. typedef Unordered_map Reloc_stub_map; // List of Cortex-A8 stubs ordered by addresses of branches being // fixed up in output. typedef std::map Cortex_a8_stub_list; // List of Arm V4BX relocation stubs ordered by associated registers. typedef std::vector Arm_v4bx_stub_list; // Owner of this stub table. Arm_input_section* owner_; // The relocation stubs. Reloc_stub_map reloc_stubs_; // The cortex_a8_stubs. Cortex_a8_stub_list cortex_a8_stubs_; // The Arm V4BX relocation stubs. Arm_v4bx_stub_list arm_v4bx_stubs_; // data size of this in the previous pass. off_t prev_data_size_; // address alignment of this in the previous pass. uint64_t prev_addralign_; }; // Arm_exidx_cantunwind class. This represents an EXIDX_CANTUNWIND entry // we add to the end of an EXIDX input section that goes into the output. class Arm_exidx_cantunwind : public Output_section_data { public: Arm_exidx_cantunwind(Relobj* relobj, unsigned int shndx) : Output_section_data(8, 4, true), relobj_(relobj), shndx_(shndx) { } // Return the object containing the section pointed by this. Relobj* relobj() const { return this->relobj_; } // Return the section index of the section pointed by this. unsigned int shndx() const { return this->shndx_; } protected: void do_write(Output_file* of) { if (parameters->target().is_big_endian()) this->do_fixed_endian_write(of); else this->do_fixed_endian_write(of); } private: // Implement do_write for a given endianity. template void inline do_fixed_endian_write(Output_file*); // The object containing the section pointed by this. Relobj* relobj_; // The section index of the section pointed by this. unsigned int shndx_; }; // During EXIDX coverage fix-up, we compact an EXIDX section. The // Offset map is used to map input section offset within the EXIDX section // to the output offset from the start of this EXIDX section. typedef std::map Arm_exidx_section_offset_map; // Arm_exidx_merged_section class. This represents an EXIDX input section // with some of its entries merged. class Arm_exidx_merged_section : public Output_relaxed_input_section { public: // Constructor for Arm_exidx_merged_section. // EXIDX_INPUT_SECTION points to the unmodified EXIDX input section. // SECTION_OFFSET_MAP points to a section offset map describing how // parts of the input section are mapped to output. DELETED_BYTES is // the number of bytes deleted from the EXIDX input section. Arm_exidx_merged_section( const Arm_exidx_input_section& exidx_input_section, const Arm_exidx_section_offset_map& section_offset_map, uint32_t deleted_bytes); // Return the original EXIDX input section. const Arm_exidx_input_section& exidx_input_section() const { return this->exidx_input_section_; } // Return the section offset map. const Arm_exidx_section_offset_map& section_offset_map() const { return this->section_offset_map_; } protected: // Write merged section into file OF. void do_write(Output_file* of); bool do_output_offset(const Relobj*, unsigned int, section_offset_type, section_offset_type*) const; private: // Original EXIDX input section. const Arm_exidx_input_section& exidx_input_section_; // Section offset map. const Arm_exidx_section_offset_map& section_offset_map_; }; // A class to wrap an ordinary input section containing executable code. template class Arm_input_section : public Output_relaxed_input_section { public: Arm_input_section(Relobj* relobj, unsigned int shndx) : Output_relaxed_input_section(relobj, shndx, 1), original_addralign_(1), original_size_(0), stub_table_(NULL) { } ~Arm_input_section() { } // Initialize. void init(); // Whether this is a stub table owner. bool is_stub_table_owner() const { return this->stub_table_ != NULL && this->stub_table_->owner() == this; } // Return the stub table. Stub_table* stub_table() const { return this->stub_table_; } // Set the stub_table. void set_stub_table(Stub_table* stub_table) { this->stub_table_ = stub_table; } // Downcast a base pointer to an Arm_input_section pointer. This is // not type-safe but we only use Arm_input_section not the base class. static Arm_input_section* as_arm_input_section(Output_relaxed_input_section* poris) { return static_cast*>(poris); } protected: // Write data to output file. void do_write(Output_file*); // Return required alignment of this. uint64_t do_addralign() const { if (this->is_stub_table_owner()) return std::max(this->stub_table_->addralign(), this->original_addralign_); else return this->original_addralign_; } // Finalize data size. void set_final_data_size(); // Reset address and file offset. void do_reset_address_and_file_offset(); // Output offset. bool do_output_offset(const Relobj* object, unsigned int shndx, section_offset_type offset, section_offset_type* poutput) const { if ((object == this->relobj()) && (shndx == this->shndx()) && (offset >= 0) && (convert_types(offset) <= this->original_size_)) { *poutput = offset; return true; } else return false; } private: // Copying is not allowed. Arm_input_section(const Arm_input_section&); Arm_input_section& operator=(const Arm_input_section&); // Address alignment of the original input section. uint64_t original_addralign_; // Section size of the original input section. uint64_t original_size_; // Stub table. Stub_table* stub_table_; }; // Arm_exidx_fixup class. This is used to define a number of methods // and keep states for fixing up EXIDX coverage. class Arm_exidx_fixup { public: Arm_exidx_fixup(Output_section* exidx_output_section) : exidx_output_section_(exidx_output_section), last_unwind_type_(UT_NONE), last_inlined_entry_(0), last_input_section_(NULL), section_offset_map_(NULL), first_output_text_section_(NULL) { } ~Arm_exidx_fixup() { delete this->section_offset_map_; } // Process an EXIDX section for entry merging. Return number of bytes to // be deleted in output. If parts of the input EXIDX section are merged // a heap allocated Arm_exidx_section_offset_map is store in the located // PSECTION_OFFSET_MAP. The caller owns the map and is reponsible for // releasing it. template uint32_t process_exidx_section(const Arm_exidx_input_section* exidx_input_section, Arm_exidx_section_offset_map** psection_offset_map); // Append an EXIDX_CANTUNWIND entry pointing at the end of the last // input section, if there is not one already. void add_exidx_cantunwind_as_needed(); // Return the output section for the text section which is linked to the // first exidx input in output. Output_section* first_output_text_section() const { return this->first_output_text_section_; } private: // Copying is not allowed. Arm_exidx_fixup(const Arm_exidx_fixup&); Arm_exidx_fixup& operator=(const Arm_exidx_fixup&); // Type of EXIDX unwind entry. enum Unwind_type { // No type. UT_NONE, // EXIDX_CANTUNWIND. UT_EXIDX_CANTUNWIND, // Inlined entry. UT_INLINED_ENTRY, // Normal entry. UT_NORMAL_ENTRY, }; // Process an EXIDX entry. We only care about the second word of the // entry. Return true if the entry can be deleted. bool process_exidx_entry(uint32_t second_word); // Update the current section offset map during EXIDX section fix-up. // If there is no map, create one. INPUT_OFFSET is the offset of a // reference point, DELETED_BYTES is the number of deleted by in the // section so far. If DELETE_ENTRY is true, the reference point and // all offsets after the previous reference point are discarded. void update_offset_map(section_offset_type input_offset, section_size_type deleted_bytes, bool delete_entry); // EXIDX output section. Output_section* exidx_output_section_; // Unwind type of the last EXIDX entry processed. Unwind_type last_unwind_type_; // Last seen inlined EXIDX entry. uint32_t last_inlined_entry_; // Last processed EXIDX input section. const Arm_exidx_input_section* last_input_section_; // Section offset map created in process_exidx_section. Arm_exidx_section_offset_map* section_offset_map_; // Output section for the text section which is linked to the first exidx // input in output. Output_section* first_output_text_section_; }; // Arm output section class. This is defined mainly to add a number of // stub generation methods. template class Arm_output_section : public Output_section { public: typedef std::vector > Text_section_list; Arm_output_section(const char* name, elfcpp::Elf_Word type, elfcpp::Elf_Xword flags) : Output_section(name, type, flags) { } ~Arm_output_section() { } // Group input sections for stub generation. void group_sections(section_size_type, bool, Target_arm*); // Downcast a base pointer to an Arm_output_section pointer. This is // not type-safe but we only use Arm_output_section not the base class. static Arm_output_section* as_arm_output_section(Output_section* os) { return static_cast*>(os); } // Append all input text sections in this into LIST. void append_text_sections_to_list(Text_section_list* list); // Fix EXIDX coverage of this EXIDX output section. SORTED_TEXT_SECTION // is a list of text input sections sorted in ascending order of their // output addresses. void fix_exidx_coverage(const Text_section_list& sorted_text_section, Symbol_table* symtab); private: // For convenience. typedef Output_section::Input_section Input_section; typedef Output_section::Input_section_list Input_section_list; // Create a stub group. void create_stub_group(Input_section_list::const_iterator, Input_section_list::const_iterator, Input_section_list::const_iterator, Target_arm*, std::vector*); }; // Arm_exidx_input_section class. This represents an EXIDX input section. class Arm_exidx_input_section { public: static const section_offset_type invalid_offset = static_cast(-1); Arm_exidx_input_section(Relobj* relobj, unsigned int shndx, unsigned int link, uint32_t size, uint32_t addralign) : relobj_(relobj), shndx_(shndx), link_(link), size_(size), addralign_(addralign) { } ~Arm_exidx_input_section() { } // Accessors: This is a read-only class. // Return the object containing this EXIDX input section. Relobj* relobj() const { return this->relobj_; } // Return the section index of this EXIDX input section. unsigned int shndx() const { return this->shndx_; } // Return the section index of linked text section in the same object. unsigned int link() const { return this->link_; } // Return size of the EXIDX input section. uint32_t size() const { return this->size_; } // Reutnr address alignment of EXIDX input section. uint32_t addralign() const { return this->addralign_; } private: // Object containing this. Relobj* relobj_; // Section index of this. unsigned int shndx_; // text section linked to this in the same object. unsigned int link_; // Size of this. For ARM 32-bit is sufficient. uint32_t size_; // Address alignment of this. For ARM 32-bit is sufficient. uint32_t addralign_; }; // Arm_relobj class. template class Arm_relobj : public Sized_relobj<32, big_endian> { public: static const Arm_address invalid_address = static_cast(-1); Arm_relobj(const std::string& name, Input_file* input_file, off_t offset, const typename elfcpp::Ehdr<32, big_endian>& ehdr) : Sized_relobj<32, big_endian>(name, input_file, offset, ehdr), stub_tables_(), local_symbol_is_thumb_function_(), attributes_section_data_(NULL), mapping_symbols_info_(), section_has_cortex_a8_workaround_(NULL), exidx_section_map_(), output_local_symbol_count_needs_update_(false) { } ~Arm_relobj() { delete this->attributes_section_data_; } // Return the stub table of the SHNDX-th section if there is one. Stub_table* stub_table(unsigned int shndx) const { gold_assert(shndx < this->stub_tables_.size()); return this->stub_tables_[shndx]; } // Set STUB_TABLE to be the stub_table of the SHNDX-th section. void set_stub_table(unsigned int shndx, Stub_table* stub_table) { gold_assert(shndx < this->stub_tables_.size()); this->stub_tables_[shndx] = stub_table; } // Whether a local symbol is a THUMB function. R_SYM is the symbol table // index. This is only valid after do_count_local_symbol is called. bool local_symbol_is_thumb_function(unsigned int r_sym) const { gold_assert(r_sym < this->local_symbol_is_thumb_function_.size()); return this->local_symbol_is_thumb_function_[r_sym]; } // Scan all relocation sections for stub generation. void scan_sections_for_stubs(Target_arm*, const Symbol_table*, const Layout*); // Convert regular input section with index SHNDX to a relaxed section. void convert_input_section_to_relaxed_section(unsigned shndx) { // The stubs have relocations and we need to process them after writing // out the stubs. So relocation now must follow section write. this->set_section_offset(shndx, -1ULL); this->set_relocs_must_follow_section_writes(); } // Downcast a base pointer to an Arm_relobj pointer. This is // not type-safe but we only use Arm_relobj not the base class. static Arm_relobj* as_arm_relobj(Relobj* relobj) { return static_cast*>(relobj); } // Processor-specific flags in ELF file header. This is valid only after // reading symbols. elfcpp::Elf_Word processor_specific_flags() const { return this->processor_specific_flags_; } // Attribute section data This is the contents of the .ARM.attribute section // if there is one. const Attributes_section_data* attributes_section_data() const { return this->attributes_section_data_; } // Mapping symbol location. typedef std::pair Mapping_symbol_position; // Functor for STL container. struct Mapping_symbol_position_less { bool operator()(const Mapping_symbol_position& p1, const Mapping_symbol_position& p2) const { return (p1.first < p2.first || (p1.first == p2.first && p1.second < p2.second)); } }; // We only care about the first character of a mapping symbol, so // we only store that instead of the whole symbol name. typedef std::map Mapping_symbols_info; // Whether a section contains any Cortex-A8 workaround. bool section_has_cortex_a8_workaround(unsigned int shndx) const { return (this->section_has_cortex_a8_workaround_ != NULL && (*this->section_has_cortex_a8_workaround_)[shndx]); } // Mark a section that has Cortex-A8 workaround. void mark_section_for_cortex_a8_workaround(unsigned int shndx) { if (this->section_has_cortex_a8_workaround_ == NULL) this->section_has_cortex_a8_workaround_ = new std::vector(this->shnum(), false); (*this->section_has_cortex_a8_workaround_)[shndx] = true; } // Return the EXIDX section of an text section with index SHNDX or NULL // if the text section has no associated EXIDX section. const Arm_exidx_input_section* exidx_input_section_by_link(unsigned int shndx) const { Exidx_section_map::const_iterator p = this->exidx_section_map_.find(shndx); return ((p != this->exidx_section_map_.end() && p->second->link() == shndx) ? p->second : NULL); } // Return the EXIDX section with index SHNDX or NULL if there is none. const Arm_exidx_input_section* exidx_input_section_by_shndx(unsigned shndx) const { Exidx_section_map::const_iterator p = this->exidx_section_map_.find(shndx); return ((p != this->exidx_section_map_.end() && p->second->shndx() == shndx) ? p->second : NULL); } // Whether output local symbol count needs updating. bool output_local_symbol_count_needs_update() const { return this->output_local_symbol_count_needs_update_; } // Set output_local_symbol_count_needs_update flag to be true. void set_output_local_symbol_count_needs_update() { this->output_local_symbol_count_needs_update_ = true; } // Update output local symbol count at the end of relaxation. void update_output_local_symbol_count(); protected: // Post constructor setup. void do_setup() { // Call parent's setup method. Sized_relobj<32, big_endian>::do_setup(); // Initialize look-up tables. Stub_table_list empty_stub_table_list(this->shnum(), NULL); this->stub_tables_.swap(empty_stub_table_list); } // Count the local symbols. void do_count_local_symbols(Stringpool_template*, Stringpool_template*); void do_relocate_sections(const Symbol_table* symtab, const Layout* layout, const unsigned char* pshdrs, typename Sized_relobj<32, big_endian>::Views* pivews); // Read the symbol information. void do_read_symbols(Read_symbols_data* sd); // Process relocs for garbage collection. void do_gc_process_relocs(Symbol_table*, Layout*, Read_relocs_data*); private: // Whether a section needs to be scanned for relocation stubs. bool section_needs_reloc_stub_scanning(const elfcpp::Shdr<32, big_endian>&, const Relobj::Output_sections&, const Symbol_table *, const unsigned char*); // Whether a section is a scannable text section. bool section_is_scannable(const elfcpp::Shdr<32, big_endian>&, unsigned int, const Output_section*, const Symbol_table *); // Whether a section needs to be scanned for the Cortex-A8 erratum. bool section_needs_cortex_a8_stub_scanning(const elfcpp::Shdr<32, big_endian>&, unsigned int, Output_section*, const Symbol_table *); // Scan a section for the Cortex-A8 erratum. void scan_section_for_cortex_a8_erratum(const elfcpp::Shdr<32, big_endian>&, unsigned int, Output_section*, Target_arm*); // Make a new Arm_exidx_input_section object for EXIDX section with // index SHNDX and section header SHDR. void make_exidx_input_section(unsigned int shndx, const elfcpp::Shdr<32, big_endian>& shdr); // Return the output address of either a plain input section or a // relaxed input section. SHNDX is the section index. Arm_address simple_input_section_output_address(unsigned int, Output_section*); typedef std::vector*> Stub_table_list; typedef Unordered_map Exidx_section_map; // List of stub tables. Stub_table_list stub_tables_; // Bit vector to tell if a local symbol is a thumb function or not. // This is only valid after do_count_local_symbol is called. std::vector local_symbol_is_thumb_function_; // processor-specific flags in ELF file header. elfcpp::Elf_Word processor_specific_flags_; // Object attributes if there is an .ARM.attributes section or NULL. Attributes_section_data* attributes_section_data_; // Mapping symbols information. Mapping_symbols_info mapping_symbols_info_; // Bitmap to indicate sections with Cortex-A8 workaround or NULL. std::vector* section_has_cortex_a8_workaround_; // Map a text section to its associated .ARM.exidx section, if there is one. Exidx_section_map exidx_section_map_; // Whether output local symbol count needs updating. bool output_local_symbol_count_needs_update_; }; // Arm_dynobj class. template class Arm_dynobj : public Sized_dynobj<32, big_endian> { public: Arm_dynobj(const std::string& name, Input_file* input_file, off_t offset, const elfcpp::Ehdr<32, big_endian>& ehdr) : Sized_dynobj<32, big_endian>(name, input_file, offset, ehdr), processor_specific_flags_(0), attributes_section_data_(NULL) { } ~Arm_dynobj() { delete this->attributes_section_data_; } // Downcast a base pointer to an Arm_relobj pointer. This is // not type-safe but we only use Arm_relobj not the base class. static Arm_dynobj* as_arm_dynobj(Dynobj* dynobj) { return static_cast*>(dynobj); } // Processor-specific flags in ELF file header. This is valid only after // reading symbols. elfcpp::Elf_Word processor_specific_flags() const { return this->processor_specific_flags_; } // Attributes section data. const Attributes_section_data* attributes_section_data() const { return this->attributes_section_data_; } protected: // Read the symbol information. void do_read_symbols(Read_symbols_data* sd); private: // processor-specific flags in ELF file header. elfcpp::Elf_Word processor_specific_flags_; // Object attributes if there is an .ARM.attributes section or NULL. Attributes_section_data* attributes_section_data_; }; // Functor to read reloc addends during stub generation. template struct Stub_addend_reader { // Return the addend for a relocation of a particular type. Depending // on whether this is a REL or RELA relocation, read the addend from a // view or from a Reloc object. elfcpp::Elf_types<32>::Elf_Swxword operator()( unsigned int /* r_type */, const unsigned char* /* view */, const typename Reloc_types::Reloc& /* reloc */) const; }; // Specialized Stub_addend_reader for SHT_REL type relocation sections. template struct Stub_addend_reader { elfcpp::Elf_types<32>::Elf_Swxword operator()( unsigned int, const unsigned char*, const typename Reloc_types::Reloc&) const; }; // Specialized Stub_addend_reader for RELA type relocation sections. // We currently do not handle RELA type relocation sections but it is trivial // to implement the addend reader. This is provided for completeness and to // make it easier to add support for RELA relocation sections in the future. template struct Stub_addend_reader { elfcpp::Elf_types<32>::Elf_Swxword operator()( unsigned int, const unsigned char*, const typename Reloc_types::Reloc& reloc) const { return reloc.get_r_addend(); } }; // Cortex_a8_reloc class. We keep record of relocation that may need // the Cortex-A8 erratum workaround. class Cortex_a8_reloc { public: Cortex_a8_reloc(Reloc_stub* reloc_stub, unsigned r_type, Arm_address destination) : reloc_stub_(reloc_stub), r_type_(r_type), destination_(destination) { } ~Cortex_a8_reloc() { } // Accessors: This is a read-only class. // Return the relocation stub associated with this relocation if there is // one. const Reloc_stub* reloc_stub() const { return this->reloc_stub_; } // Return the relocation type. unsigned int r_type() const { return this->r_type_; } // Return the destination address of the relocation. LSB stores the THUMB // bit. Arm_address destination() const { return this->destination_; } private: // Associated relocation stub if there is one, or NULL. const Reloc_stub* reloc_stub_; // Relocation type. unsigned int r_type_; // Destination address of this relocation. LSB is used to distinguish // ARM/THUMB mode. Arm_address destination_; }; // Utilities for manipulating integers of up to 32-bits namespace utils { // Sign extend an n-bit unsigned integer stored in an uint32_t into // an int32_t. NO_BITS must be between 1 to 32. template static inline int32_t sign_extend(uint32_t bits) { gold_assert(no_bits >= 0 && no_bits <= 32); if (no_bits == 32) return static_cast(bits); uint32_t mask = (~((uint32_t) 0)) >> (32 - no_bits); bits &= mask; uint32_t top_bit = 1U << (no_bits - 1); int32_t as_signed = static_cast(bits); return (bits & top_bit) ? as_signed + (-top_bit * 2) : as_signed; } // Detects overflow of an NO_BITS integer stored in a uint32_t. template static inline bool has_overflow(uint32_t bits) { gold_assert(no_bits >= 0 && no_bits <= 32); if (no_bits == 32) return false; int32_t max = (1 << (no_bits - 1)) - 1; int32_t min = -(1 << (no_bits - 1)); int32_t as_signed = static_cast(bits); return as_signed > max || as_signed < min; } // Detects overflow of an NO_BITS integer stored in a uint32_t when it // fits in the given number of bits as either a signed or unsigned value. // For example, has_signed_unsigned_overflow<8> would check // -128 <= bits <= 255 template static inline bool has_signed_unsigned_overflow(uint32_t bits) { gold_assert(no_bits >= 2 && no_bits <= 32); if (no_bits == 32) return false; int32_t max = static_cast((1U << no_bits) - 1); int32_t min = -(1 << (no_bits - 1)); int32_t as_signed = static_cast(bits); return as_signed > max || as_signed < min; } // Select bits from A and B using bits in MASK. For each n in [0..31], // the n-th bit in the result is chosen from the n-th bits of A and B. // A zero selects A and a one selects B. static inline uint32_t bit_select(uint32_t a, uint32_t b, uint32_t mask) { return (a & ~mask) | (b & mask); } }; template class Target_arm : public Sized_target<32, big_endian> { public: typedef Output_data_reloc Reloc_section; // When were are relocating a stub, we pass this as the relocation number. static const size_t fake_relnum_for_stubs = static_cast(-1); Target_arm() : Sized_target<32, big_endian>(&arm_info), got_(NULL), plt_(NULL), got_plt_(NULL), rel_dyn_(NULL), copy_relocs_(elfcpp::R_ARM_COPY), dynbss_(NULL), stub_tables_(), stub_factory_(Stub_factory::get_instance()), may_use_blx_(false), should_force_pic_veneer_(false), arm_input_section_map_(), attributes_section_data_(NULL), fix_cortex_a8_(false), cortex_a8_relocs_info_() { } // Whether we can use BLX. bool may_use_blx() const { return this->may_use_blx_; } // Set use-BLX flag. void set_may_use_blx(bool value) { this->may_use_blx_ = value; } // Whether we force PCI branch veneers. bool should_force_pic_veneer() const { return this->should_force_pic_veneer_; } // Set PIC veneer flag. void set_should_force_pic_veneer(bool value) { this->should_force_pic_veneer_ = value; } // Whether we use THUMB-2 instructions. bool using_thumb2() const { Object_attribute* attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch); int arch = attr->int_value(); return arch == elfcpp::TAG_CPU_ARCH_V6T2 || arch >= elfcpp::TAG_CPU_ARCH_V7; } // Whether we use THUMB/THUMB-2 instructions only. bool using_thumb_only() const { Object_attribute* attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch); if (attr->int_value() != elfcpp::TAG_CPU_ARCH_V7 && attr->int_value() != elfcpp::TAG_CPU_ARCH_V7E_M) return false; attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch_profile); return attr->int_value() == 'M'; } // Whether we have an NOP instruction. If not, use mov r0, r0 instead. bool may_use_arm_nop() const { Object_attribute* attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch); int arch = attr->int_value(); return (arch == elfcpp::TAG_CPU_ARCH_V6T2 || arch == elfcpp::TAG_CPU_ARCH_V6K || arch == elfcpp::TAG_CPU_ARCH_V7 || arch == elfcpp::TAG_CPU_ARCH_V7E_M); } // Whether we have THUMB-2 NOP.W instruction. bool may_use_thumb2_nop() const { Object_attribute* attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch); int arch = attr->int_value(); return (arch == elfcpp::TAG_CPU_ARCH_V6T2 || arch == elfcpp::TAG_CPU_ARCH_V7 || arch == elfcpp::TAG_CPU_ARCH_V7E_M); } // Process the relocations to determine unreferenced sections for // garbage collection. void gc_process_relocs(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols); // Scan the relocations to look for symbol adjustments. void scan_relocs(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols); // Finalize the sections. void do_finalize_sections(Layout*, const Input_objects*, Symbol_table*); // Return the value to use for a dynamic symbol which requires special // treatment. uint64_t do_dynsym_value(const Symbol*) const; // Relocate a section. void relocate_section(const Relocate_info<32, big_endian>*, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, unsigned char* view, Arm_address view_address, section_size_type view_size, const Reloc_symbol_changes*); // Scan the relocs during a relocatable link. void scan_relocatable_relocs(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols, Relocatable_relocs*); // Relocate a section during a relocatable link. void relocate_for_relocatable(const Relocate_info<32, big_endian>*, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, off_t offset_in_output_section, const Relocatable_relocs*, unsigned char* view, Arm_address view_address, section_size_type view_size, unsigned char* reloc_view, section_size_type reloc_view_size); // Return whether SYM is defined by the ABI. bool do_is_defined_by_abi(Symbol* sym) const { return strcmp(sym->name(), "__tls_get_addr") == 0; } // Return the size of the GOT section. section_size_type got_size() { gold_assert(this->got_ != NULL); return this->got_->data_size(); } // Map platform-specific reloc types static unsigned int get_real_reloc_type (unsigned int r_type); // // Methods to support stub-generations. // // Return the stub factory const Stub_factory& stub_factory() const { return this->stub_factory_; } // Make a new Arm_input_section object. Arm_input_section* new_arm_input_section(Relobj*, unsigned int); // Find the Arm_input_section object corresponding to the SHNDX-th input // section of RELOBJ. Arm_input_section* find_arm_input_section(Relobj* relobj, unsigned int shndx) const; // Make a new Stub_table Stub_table* new_stub_table(Arm_input_section*); // Scan a section for stub generation. void scan_section_for_stubs(const Relocate_info<32, big_endian>*, unsigned int, const unsigned char*, size_t, Output_section*, bool, const unsigned char*, Arm_address, section_size_type); // Relocate a stub. void relocate_stub(Stub*, const Relocate_info<32, big_endian>*, Output_section*, unsigned char*, Arm_address, section_size_type); // Get the default ARM target. static Target_arm* default_target() { gold_assert(parameters->target().machine_code() == elfcpp::EM_ARM && parameters->target().is_big_endian() == big_endian); return static_cast*>( parameters->sized_target<32, big_endian>()); } // Whether NAME belongs to a mapping symbol. static bool is_mapping_symbol_name(const char* name) { return (name && name[0] == '$' && (name[1] == 'a' || name[1] == 't' || name[1] == 'd') && (name[2] == '\0' || name[2] == '.')); } // Whether we work around the Cortex-A8 erratum. bool fix_cortex_a8() const { return this->fix_cortex_a8_; } // Whether we fix R_ARM_V4BX relocation. // 0 - do not fix // 1 - replace with MOV instruction (armv4 target) // 2 - make interworking veneer (>= armv4t targets only) General_options::Fix_v4bx fix_v4bx() const { return parameters->options().fix_v4bx(); } // Scan a span of THUMB code section for Cortex-A8 erratum. void scan_span_for_cortex_a8_erratum(Arm_relobj*, unsigned int, section_size_type, section_size_type, const unsigned char*, Arm_address); // Apply Cortex-A8 workaround to a branch. void apply_cortex_a8_workaround(const Cortex_a8_stub*, Arm_address, unsigned char*, Arm_address); protected: // Make an ELF object. Object* do_make_elf_object(const std::string&, Input_file*, off_t, const elfcpp::Ehdr<32, big_endian>& ehdr); Object* do_make_elf_object(const std::string&, Input_file*, off_t, const elfcpp::Ehdr<32, !big_endian>&) { gold_unreachable(); } Object* do_make_elf_object(const std::string&, Input_file*, off_t, const elfcpp::Ehdr<64, false>&) { gold_unreachable(); } Object* do_make_elf_object(const std::string&, Input_file*, off_t, const elfcpp::Ehdr<64, true>&) { gold_unreachable(); } // Make an output section. Output_section* do_make_output_section(const char* name, elfcpp::Elf_Word type, elfcpp::Elf_Xword flags) { return new Arm_output_section(name, type, flags); } void do_adjust_elf_header(unsigned char* view, int len) const; // We only need to generate stubs, and hence perform relaxation if we are // not doing relocatable linking. bool do_may_relax() const { return !parameters->options().relocatable(); } bool do_relax(int, const Input_objects*, Symbol_table*, Layout*); // Determine whether an object attribute tag takes an integer, a // string or both. int do_attribute_arg_type(int tag) const; // Reorder tags during output. int do_attributes_order(int num) const; // This is called when the target is selected as the default. void do_select_as_default_target() { // No locking is required since there should only be one default target. // We cannot have both the big-endian and little-endian ARM targets // as the default. gold_assert(arm_reloc_property_table == NULL); arm_reloc_property_table = new Arm_reloc_property_table(); } private: // The class which scans relocations. class Scan { public: Scan() : issued_non_pic_error_(false) { } inline void local(Symbol_table* symtab, Layout* layout, Target_arm* target, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, Output_section* output_section, const elfcpp::Rel<32, big_endian>& reloc, unsigned int r_type, const elfcpp::Sym<32, big_endian>& lsym); inline void global(Symbol_table* symtab, Layout* layout, Target_arm* target, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, Output_section* output_section, const elfcpp::Rel<32, big_endian>& reloc, unsigned int r_type, Symbol* gsym); private: static void unsupported_reloc_local(Sized_relobj<32, big_endian>*, unsigned int r_type); static void unsupported_reloc_global(Sized_relobj<32, big_endian>*, unsigned int r_type, Symbol*); void check_non_pic(Relobj*, unsigned int r_type); // Almost identical to Symbol::needs_plt_entry except that it also // handles STT_ARM_TFUNC. static bool symbol_needs_plt_entry(const Symbol* sym) { // An undefined symbol from an executable does not need a PLT entry. if (sym->is_undefined() && !parameters->options().shared()) return false; return (!parameters->doing_static_link() && (sym->type() == elfcpp::STT_FUNC || sym->type() == elfcpp::STT_ARM_TFUNC) && (sym->is_from_dynobj() || sym->is_undefined() || sym->is_preemptible())); } // Whether we have issued an error about a non-PIC compilation. bool issued_non_pic_error_; }; // The class which implements relocation. class Relocate { public: Relocate() { } ~Relocate() { } // Return whether the static relocation needs to be applied. inline bool should_apply_static_reloc(const Sized_symbol<32>* gsym, int ref_flags, bool is_32bit, Output_section* output_section); // Do a relocation. Return false if the caller should not issue // any warnings about this relocation. inline bool relocate(const Relocate_info<32, big_endian>*, Target_arm*, Output_section*, size_t relnum, const elfcpp::Rel<32, big_endian>&, unsigned int r_type, const Sized_symbol<32>*, const Symbol_value<32>*, unsigned char*, Arm_address, section_size_type); // Return whether we want to pass flag NON_PIC_REF for this // reloc. This means the relocation type accesses a symbol not via // GOT or PLT. static inline bool reloc_is_non_pic (unsigned int r_type) { switch (r_type) { // These relocation types reference GOT or PLT entries explicitly. case elfcpp::R_ARM_GOT_BREL: case elfcpp::R_ARM_GOT_ABS: case elfcpp::R_ARM_GOT_PREL: case elfcpp::R_ARM_GOT_BREL12: case elfcpp::R_ARM_PLT32_ABS: case elfcpp::R_ARM_TLS_GD32: case elfcpp::R_ARM_TLS_LDM32: case elfcpp::R_ARM_TLS_IE32: case elfcpp::R_ARM_TLS_IE12GP: // These relocate types may use PLT entries. case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_THM_JUMP24: case elfcpp::R_ARM_THM_JUMP19: case elfcpp::R_ARM_PLT32: case elfcpp::R_ARM_THM_XPC22: return false; default: return true; } } }; // A class which returns the size required for a relocation type, // used while scanning relocs during a relocatable link. class Relocatable_size_for_reloc { public: unsigned int get_size_for_reloc(unsigned int, Relobj*); }; // Get the GOT section, creating it if necessary. Output_data_got<32, big_endian>* got_section(Symbol_table*, Layout*); // Get the GOT PLT section. Output_data_space* got_plt_section() const { gold_assert(this->got_plt_ != NULL); return this->got_plt_; } // Create a PLT entry for a global symbol. void make_plt_entry(Symbol_table*, Layout*, Symbol*); // Get the PLT section. const Output_data_plt_arm* plt_section() const { gold_assert(this->plt_ != NULL); return this->plt_; } // Get the dynamic reloc section, creating it if necessary. Reloc_section* rel_dyn_section(Layout*); // Return true if the symbol may need a COPY relocation. // References from an executable object to non-function symbols // defined in a dynamic object may need a COPY relocation. bool may_need_copy_reloc(Symbol* gsym) { return (gsym->type() != elfcpp::STT_ARM_TFUNC && gsym->may_need_copy_reloc()); } // Add a potential copy relocation. void copy_reloc(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int shndx, Output_section* output_section, Symbol* sym, const elfcpp::Rel<32, big_endian>& reloc) { this->copy_relocs_.copy_reloc(symtab, layout, symtab->get_sized_symbol<32>(sym), object, shndx, output_section, reloc, this->rel_dyn_section(layout)); } // Whether two EABI versions are compatible. static bool are_eabi_versions_compatible(elfcpp::Elf_Word v1, elfcpp::Elf_Word v2); // Merge processor-specific flags from input object and those in the ELF // header of the output. void merge_processor_specific_flags(const std::string&, elfcpp::Elf_Word); // Get the secondary compatible architecture. static int get_secondary_compatible_arch(const Attributes_section_data*); // Set the secondary compatible architecture. static void set_secondary_compatible_arch(Attributes_section_data*, int); static int tag_cpu_arch_combine(const char*, int, int*, int, int); // Helper to print AEABI enum tag value. static std::string aeabi_enum_name(unsigned int); // Return string value for TAG_CPU_name. static std::string tag_cpu_name_value(unsigned int); // Merge object attributes from input object and those in the output. void merge_object_attributes(const char*, const Attributes_section_data*); // Helper to get an AEABI object attribute Object_attribute* get_aeabi_object_attribute(int tag) const { Attributes_section_data* pasd = this->attributes_section_data_; gold_assert(pasd != NULL); Object_attribute* attr = pasd->get_attribute(Object_attribute::OBJ_ATTR_PROC, tag); gold_assert(attr != NULL); return attr; } // // Methods to support stub-generations. // // Group input sections for stub generation. void group_sections(Layout*, section_size_type, bool); // Scan a relocation for stub generation. void scan_reloc_for_stub(const Relocate_info<32, big_endian>*, unsigned int, const Sized_symbol<32>*, unsigned int, const Symbol_value<32>*, elfcpp::Elf_types<32>::Elf_Swxword, Arm_address); // Scan a relocation section for stub. template void scan_reloc_section_for_stubs( const Relocate_info<32, big_endian>* relinfo, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, const unsigned char* view, elfcpp::Elf_types<32>::Elf_Addr view_address, section_size_type); // Fix .ARM.exidx section coverage. void fix_exidx_coverage(Layout*, Arm_output_section*, Symbol_table*); // Functors for STL set. struct output_section_address_less_than { bool operator()(const Output_section* s1, const Output_section* s2) const { return s1->address() < s2->address(); } }; // Information about this specific target which we pass to the // general Target structure. static const Target::Target_info arm_info; // The types of GOT entries needed for this platform. enum Got_type { GOT_TYPE_STANDARD = 0 // GOT entry for a regular symbol }; typedef typename std::vector*> Stub_table_list; // Map input section to Arm_input_section. typedef Unordered_map*, Section_id_hash> Arm_input_section_map; // Map output addresses to relocs for Cortex-A8 erratum. typedef Unordered_map Cortex_a8_relocs_info; // The GOT section. Output_data_got<32, big_endian>* got_; // The PLT section. Output_data_plt_arm* plt_; // The GOT PLT section. Output_data_space* got_plt_; // The dynamic reloc section. Reloc_section* rel_dyn_; // Relocs saved to avoid a COPY reloc. Copy_relocs copy_relocs_; // Space for variables copied with a COPY reloc. Output_data_space* dynbss_; // Vector of Stub_tables created. Stub_table_list stub_tables_; // Stub factory. const Stub_factory &stub_factory_; // Whether we can use BLX. bool may_use_blx_; // Whether we force PIC branch veneers. bool should_force_pic_veneer_; // Map for locating Arm_input_sections. Arm_input_section_map arm_input_section_map_; // Attributes section data in output. Attributes_section_data* attributes_section_data_; // Whether we want to fix code for Cortex-A8 erratum. bool fix_cortex_a8_; // Map addresses to relocs for Cortex-A8 erratum. Cortex_a8_relocs_info cortex_a8_relocs_info_; }; template const Target::Target_info Target_arm::arm_info = { 32, // size big_endian, // is_big_endian elfcpp::EM_ARM, // machine_code false, // has_make_symbol false, // has_resolve false, // has_code_fill true, // is_default_stack_executable '\0', // wrap_char "/usr/lib/libc.so.1", // dynamic_linker 0x8000, // default_text_segment_address 0x1000, // abi_pagesize (overridable by -z max-page-size) 0x1000, // common_pagesize (overridable by -z common-page-size) elfcpp::SHN_UNDEF, // small_common_shndx elfcpp::SHN_UNDEF, // large_common_shndx 0, // small_common_section_flags 0, // large_common_section_flags ".ARM.attributes", // attributes_section "aeabi" // attributes_vendor }; // Arm relocate functions class // template class Arm_relocate_functions : public Relocate_functions<32, big_endian> { public: typedef enum { STATUS_OKAY, // No error during relocation. STATUS_OVERFLOW, // Relocation oveflow. STATUS_BAD_RELOC // Relocation cannot be applied. } Status; private: typedef Relocate_functions<32, big_endian> Base; typedef Arm_relocate_functions This; // Encoding of imm16 argument for movt and movw ARM instructions // from ARM ARM: // // imm16 := imm4 | imm12 // // f e d c b a 9 8 7 6 5 4 3 2 1 0 f e d c b a 9 8 7 6 5 4 3 2 1 0 // +-------+---------------+-------+-------+-----------------------+ // | | |imm4 | |imm12 | // +-------+---------------+-------+-------+-----------------------+ // Extract the relocation addend from VAL based on the ARM // instruction encoding described above. static inline typename elfcpp::Swap<32, big_endian>::Valtype extract_arm_movw_movt_addend( typename elfcpp::Swap<32, big_endian>::Valtype val) { // According to the Elf ABI for ARM Architecture the immediate // field is sign-extended to form the addend. return utils::sign_extend<16>(((val >> 4) & 0xf000) | (val & 0xfff)); } // Insert X into VAL based on the ARM instruction encoding described // above. static inline typename elfcpp::Swap<32, big_endian>::Valtype insert_val_arm_movw_movt( typename elfcpp::Swap<32, big_endian>::Valtype val, typename elfcpp::Swap<32, big_endian>::Valtype x) { val &= 0xfff0f000; val |= x & 0x0fff; val |= (x & 0xf000) << 4; return val; } // Encoding of imm16 argument for movt and movw Thumb2 instructions // from ARM ARM: // // imm16 := imm4 | i | imm3 | imm8 // // f e d c b a 9 8 7 6 5 4 3 2 1 0 f e d c b a 9 8 7 6 5 4 3 2 1 0 // +---------+-+-----------+-------++-+-----+-------+---------------+ // | |i| |imm4 || |imm3 | |imm8 | // +---------+-+-----------+-------++-+-----+-------+---------------+ // Extract the relocation addend from VAL based on the Thumb2 // instruction encoding described above. static inline typename elfcpp::Swap<32, big_endian>::Valtype extract_thumb_movw_movt_addend( typename elfcpp::Swap<32, big_endian>::Valtype val) { // According to the Elf ABI for ARM Architecture the immediate // field is sign-extended to form the addend. return utils::sign_extend<16>(((val >> 4) & 0xf000) | ((val >> 15) & 0x0800) | ((val >> 4) & 0x0700) | (val & 0x00ff)); } // Insert X into VAL based on the Thumb2 instruction encoding // described above. static inline typename elfcpp::Swap<32, big_endian>::Valtype insert_val_thumb_movw_movt( typename elfcpp::Swap<32, big_endian>::Valtype val, typename elfcpp::Swap<32, big_endian>::Valtype x) { val &= 0xfbf08f00; val |= (x & 0xf000) << 4; val |= (x & 0x0800) << 15; val |= (x & 0x0700) << 4; val |= (x & 0x00ff); return val; } // Calculate the smallest constant Kn for the specified residual. // (see (AAELF 4.6.1.4 Static ARM relocations, Group Relocations, p.32) static uint32_t calc_grp_kn(typename elfcpp::Swap<32, big_endian>::Valtype residual) { int32_t msb; if (residual == 0) return 0; // Determine the most significant bit in the residual and // align the resulting value to a 2-bit boundary. for (msb = 30; (msb >= 0) && !(residual & (3 << msb)); msb -= 2) ; // The desired shift is now (msb - 6), or zero, whichever // is the greater. return (((msb - 6) < 0) ? 0 : (msb - 6)); } // Calculate the final residual for the specified group index. // If the passed group index is less than zero, the method will return // the value of the specified residual without any change. // (see (AAELF 4.6.1.4 Static ARM relocations, Group Relocations, p.32) static typename elfcpp::Swap<32, big_endian>::Valtype calc_grp_residual(typename elfcpp::Swap<32, big_endian>::Valtype residual, const int group) { for (int n = 0; n <= group; n++) { // Calculate which part of the value to mask. uint32_t shift = calc_grp_kn(residual); // Calculate the residual for the next time around. residual &= ~(residual & (0xff << shift)); } return residual; } // Calculate the value of Gn for the specified group index. // We return it in the form of an encoded constant-and-rotation. // (see (AAELF 4.6.1.4 Static ARM relocations, Group Relocations, p.32) static typename elfcpp::Swap<32, big_endian>::Valtype calc_grp_gn(typename elfcpp::Swap<32, big_endian>::Valtype residual, const int group) { typename elfcpp::Swap<32, big_endian>::Valtype gn = 0; uint32_t shift = 0; for (int n = 0; n <= group; n++) { // Calculate which part of the value to mask. shift = calc_grp_kn(residual); // Calculate Gn in 32-bit as well as encoded constant-and-rotation form. gn = residual & (0xff << shift); // Calculate the residual for the next time around. residual &= ~gn; } // Return Gn in the form of an encoded constant-and-rotation. return ((gn >> shift) | ((gn <= 0xff ? 0 : (32 - shift) / 2) << 8)); } public: // Handle ARM long branches. static typename This::Status arm_branch_common(unsigned int, const Relocate_info<32, big_endian>*, unsigned char *, const Sized_symbol<32>*, const Arm_relobj*, unsigned int, const Symbol_value<32>*, Arm_address, Arm_address, bool); // Handle THUMB long branches. static typename This::Status thumb_branch_common(unsigned int, const Relocate_info<32, big_endian>*, unsigned char *, const Sized_symbol<32>*, const Arm_relobj*, unsigned int, const Symbol_value<32>*, Arm_address, Arm_address, bool); // Return the branch offset of a 32-bit THUMB branch. static inline int32_t thumb32_branch_offset(uint16_t upper_insn, uint16_t lower_insn) { // We use the Thumb-2 encoding (backwards compatible with Thumb-1) // involving the J1 and J2 bits. uint32_t s = (upper_insn & (1U << 10)) >> 10; uint32_t upper = upper_insn & 0x3ffU; uint32_t lower = lower_insn & 0x7ffU; uint32_t j1 = (lower_insn & (1U << 13)) >> 13; uint32_t j2 = (lower_insn & (1U << 11)) >> 11; uint32_t i1 = j1 ^ s ? 0 : 1; uint32_t i2 = j2 ^ s ? 0 : 1; return utils::sign_extend<25>((s << 24) | (i1 << 23) | (i2 << 22) | (upper << 12) | (lower << 1)); } // Insert OFFSET to a 32-bit THUMB branch and return the upper instruction. // UPPER_INSN is the original upper instruction of the branch. Caller is // responsible for overflow checking and BLX offset adjustment. static inline uint16_t thumb32_branch_upper(uint16_t upper_insn, int32_t offset) { uint32_t s = offset < 0 ? 1 : 0; uint32_t bits = static_cast(offset); return (upper_insn & ~0x7ffU) | ((bits >> 12) & 0x3ffU) | (s << 10); } // Insert OFFSET to a 32-bit THUMB branch and return the lower instruction. // LOWER_INSN is the original lower instruction of the branch. Caller is // responsible for overflow checking and BLX offset adjustment. static inline uint16_t thumb32_branch_lower(uint16_t lower_insn, int32_t offset) { uint32_t s = offset < 0 ? 1 : 0; uint32_t bits = static_cast(offset); return ((lower_insn & ~0x2fffU) | ((((bits >> 23) & 1) ^ !s) << 13) | ((((bits >> 22) & 1) ^ !s) << 11) | ((bits >> 1) & 0x7ffU)); } // Return the branch offset of a 32-bit THUMB conditional branch. static inline int32_t thumb32_cond_branch_offset(uint16_t upper_insn, uint16_t lower_insn) { uint32_t s = (upper_insn & 0x0400U) >> 10; uint32_t j1 = (lower_insn & 0x2000U) >> 13; uint32_t j2 = (lower_insn & 0x0800U) >> 11; uint32_t lower = (lower_insn & 0x07ffU); uint32_t upper = (s << 8) | (j2 << 7) | (j1 << 6) | (upper_insn & 0x003fU); return utils::sign_extend<21>((upper << 12) | (lower << 1)); } // Insert OFFSET to a 32-bit THUMB conditional branch and return the upper // instruction. UPPER_INSN is the original upper instruction of the branch. // Caller is responsible for overflow checking. static inline uint16_t thumb32_cond_branch_upper(uint16_t upper_insn, int32_t offset) { uint32_t s = offset < 0 ? 1 : 0; uint32_t bits = static_cast(offset); return (upper_insn & 0xfbc0U) | (s << 10) | ((bits & 0x0003f000U) >> 12); } // Insert OFFSET to a 32-bit THUMB conditional branch and return the lower // instruction. LOWER_INSN is the original lower instruction of the branch. // Caller is reponsible for overflow checking. static inline uint16_t thumb32_cond_branch_lower(uint16_t lower_insn, int32_t offset) { uint32_t bits = static_cast(offset); uint32_t j2 = (bits & 0x00080000U) >> 19; uint32_t j1 = (bits & 0x00040000U) >> 18; uint32_t lo = (bits & 0x00000ffeU) >> 1; return (lower_insn & 0xd000U) | (j1 << 13) | (j2 << 11) | lo; } // R_ARM_ABS8: S + A static inline typename This::Status abs8(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval) { typedef typename elfcpp::Swap<8, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<8, big_endian>::readval(wv); Reltype addend = utils::sign_extend<8>(val); Reltype x = psymval->value(object, addend); val = utils::bit_select(val, x, 0xffU); elfcpp::Swap<8, big_endian>::writeval(wv, val); return (utils::has_signed_unsigned_overflow<8>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_ABS5: S + A static inline typename This::Status thm_abs5(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<16, big_endian>::readval(wv); Reltype addend = (val & 0x7e0U) >> 6; Reltype x = psymval->value(object, addend); val = utils::bit_select(val, x << 6, 0x7e0U); elfcpp::Swap<16, big_endian>::writeval(wv, val); return (utils::has_overflow<5>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_ABS12: S + A static inline typename This::Status abs12(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); Reltype addend = val & 0x0fffU; Reltype x = psymval->value(object, addend); val = utils::bit_select(val, x, 0x0fffU); elfcpp::Swap<32, big_endian>::writeval(wv, val); return (utils::has_overflow<12>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_ABS16: S + A static inline typename This::Status abs16(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<16, big_endian>::readval(wv); Reltype addend = utils::sign_extend<16>(val); Reltype x = psymval->value(object, addend); val = utils::bit_select(val, x, 0xffffU); elfcpp::Swap<16, big_endian>::writeval(wv, val); return (utils::has_signed_unsigned_overflow<16>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_ABS32: (S + A) | T static inline typename This::Status abs32(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address thumb_bit) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype addend = elfcpp::Swap<32, big_endian>::readval(wv); Valtype x = psymval->value(object, addend) | thumb_bit; elfcpp::Swap<32, big_endian>::writeval(wv, x); return This::STATUS_OKAY; } // R_ARM_REL32: (S + A) | T - P static inline typename This::Status rel32(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype addend = elfcpp::Swap<32, big_endian>::readval(wv); Valtype x = (psymval->value(object, addend) | thumb_bit) - address; elfcpp::Swap<32, big_endian>::writeval(wv, x); return This::STATUS_OKAY; } // R_ARM_THM_JUMP24: (S + A) | T - P static typename This::Status thm_jump19(unsigned char *view, const Arm_relobj* object, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit); // R_ARM_THM_JUMP6: S + A – P static inline typename This::Status thm_jump6(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<16, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<16, big_endian>::readval(wv); // bit[9]:bit[7:3]:’0’ (mask: 0x02f8) Reltype addend = (((val & 0x0200) >> 3) | ((val & 0x00f8) >> 2)); Reltype x = (psymval->value(object, addend) - address); val = (val & 0xfd07) | ((x & 0x0040) << 3) | ((val & 0x003e) << 2); elfcpp::Swap<16, big_endian>::writeval(wv, val); // CZB does only forward jumps. return ((x > 0x007e) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_JUMP8: S + A – P static inline typename This::Status thm_jump8(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<16, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<16, big_endian>::readval(wv); Reltype addend = utils::sign_extend<8>((val & 0x00ff) << 1); Reltype x = (psymval->value(object, addend) - address); elfcpp::Swap<16, big_endian>::writeval(wv, (val & 0xff00) | ((x & 0x01fe) >> 1)); return (utils::has_overflow<8>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_JUMP11: S + A – P static inline typename This::Status thm_jump11(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<16, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<16, big_endian>::readval(wv); Reltype addend = utils::sign_extend<11>((val & 0x07ff) << 1); Reltype x = (psymval->value(object, addend) - address); elfcpp::Swap<16, big_endian>::writeval(wv, (val & 0xf800) | ((x & 0x0ffe) >> 1)); return (utils::has_overflow<11>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_BASE_PREL: B(S) + A - P static inline typename This::Status base_prel(unsigned char* view, Arm_address origin, Arm_address address) { Base::rel32(view, origin - address); return STATUS_OKAY; } // R_ARM_BASE_ABS: B(S) + A static inline typename This::Status base_abs(unsigned char* view, Arm_address origin) { Base::rel32(view, origin); return STATUS_OKAY; } // R_ARM_GOT_BREL: GOT(S) + A - GOT_ORG static inline typename This::Status got_brel(unsigned char* view, typename elfcpp::Swap<32, big_endian>::Valtype got_offset) { Base::rel32(view, got_offset); return This::STATUS_OKAY; } // R_ARM_GOT_PREL: GOT(S) + A - P static inline typename This::Status got_prel(unsigned char *view, Arm_address got_entry, Arm_address address) { Base::rel32(view, got_entry - address); return This::STATUS_OKAY; } // R_ARM_PREL: (S + A) | T - P static inline typename This::Status prel31(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); Valtype addend = utils::sign_extend<31>(val); Valtype x = (psymval->value(object, addend) | thumb_bit) - address; val = utils::bit_select(val, x, 0x7fffffffU); elfcpp::Swap<32, big_endian>::writeval(wv, val); return (utils::has_overflow<31>(x) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_MOVW_ABS_NC: (S + A) | T (relative address base is ) // R_ARM_MOVW_PREL_NC: (S + A) | T - P // R_ARM_MOVW_BREL_NC: ((S + A) | T) - B(S) // R_ARM_MOVW_BREL: ((S + A) | T) - B(S) static inline typename This::Status movw(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address relative_address_base, Arm_address thumb_bit, bool check_overflow) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); Valtype addend = This::extract_arm_movw_movt_addend(val); Valtype x = ((psymval->value(object, addend) | thumb_bit) - relative_address_base); val = This::insert_val_arm_movw_movt(val, x); elfcpp::Swap<32, big_endian>::writeval(wv, val); return ((check_overflow && utils::has_overflow<16>(x)) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_MOVT_ABS: S + A (relative address base is 0) // R_ARM_MOVT_PREL: S + A - P // R_ARM_MOVT_BREL: S + A - B(S) static inline typename This::Status movt(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address relative_address_base) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); Valtype addend = This::extract_arm_movw_movt_addend(val); Valtype x = (psymval->value(object, addend) - relative_address_base) >> 16; val = This::insert_val_arm_movw_movt(val, x); elfcpp::Swap<32, big_endian>::writeval(wv, val); // FIXME: IHI0044D says that we should check for overflow. return This::STATUS_OKAY; } // R_ARM_THM_MOVW_ABS_NC: S + A | T (relative_address_base is 0) // R_ARM_THM_MOVW_PREL_NC: (S + A) | T - P // R_ARM_THM_MOVW_BREL_NC: ((S + A) | T) - B(S) // R_ARM_THM_MOVW_BREL: ((S + A) | T) - B(S) static inline typename This::Status thm_movw(unsigned char *view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address relative_address_base, Arm_address thumb_bit, bool check_overflow) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Reltype val = (elfcpp::Swap<16, big_endian>::readval(wv) << 16) | elfcpp::Swap<16, big_endian>::readval(wv + 1); Reltype addend = This::extract_thumb_movw_movt_addend(val); Reltype x = (psymval->value(object, addend) | thumb_bit) - relative_address_base; val = This::insert_val_thumb_movw_movt(val, x); elfcpp::Swap<16, big_endian>::writeval(wv, val >> 16); elfcpp::Swap<16, big_endian>::writeval(wv + 1, val & 0xffff); return ((check_overflow && utils::has_overflow<16>(x)) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_MOVT_ABS: S + A (relative address base is 0) // R_ARM_THM_MOVT_PREL: S + A - P // R_ARM_THM_MOVT_BREL: S + A - B(S) static inline typename This::Status thm_movt(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address relative_address_base) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Reltype val = (elfcpp::Swap<16, big_endian>::readval(wv) << 16) | elfcpp::Swap<16, big_endian>::readval(wv + 1); Reltype addend = This::extract_thumb_movw_movt_addend(val); Reltype x = (psymval->value(object, addend) - relative_address_base) >> 16; val = This::insert_val_thumb_movw_movt(val, x); elfcpp::Swap<16, big_endian>::writeval(wv, val >> 16); elfcpp::Swap<16, big_endian>::writeval(wv + 1, val & 0xffff); return This::STATUS_OKAY; } // R_ARM_THM_ALU_PREL_11_0: ((S + A) | T) - Pa (Thumb32) static inline typename This::Status thm_alu11(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Reltype insn = (elfcpp::Swap<16, big_endian>::readval(wv) << 16) | elfcpp::Swap<16, big_endian>::readval(wv + 1); // f e d c b|a|9|8 7 6 5|4|3 2 1 0||f|e d c|b a 9 8|7 6 5 4 3 2 1 0 // ----------------------------------------------------------------------- // ADD{S} 1 1 1 1 0|i|0|1 0 0 0|S|1 1 0 1||0|imm3 |Rd |imm8 // ADDW 1 1 1 1 0|i|1|0 0 0 0|0|1 1 0 1||0|imm3 |Rd |imm8 // ADR[+] 1 1 1 1 0|i|1|0 0 0 0|0|1 1 1 1||0|imm3 |Rd |imm8 // SUB{S} 1 1 1 1 0|i|0|1 1 0 1|S|1 1 0 1||0|imm3 |Rd |imm8 // SUBW 1 1 1 1 0|i|1|0 1 0 1|0|1 1 0 1||0|imm3 |Rd |imm8 // ADR[-] 1 1 1 1 0|i|1|0 1 0 1|0|1 1 1 1||0|imm3 |Rd |imm8 // Determine a sign for the addend. const int sign = ((insn & 0xf8ef0000) == 0xf0ad0000 || (insn & 0xf8ef0000) == 0xf0af0000) ? -1 : 1; // Thumb2 addend encoding: // imm12 := i | imm3 | imm8 int32_t addend = (insn & 0xff) | ((insn & 0x00007000) >> 4) | ((insn & 0x04000000) >> 15); // Apply a sign to the added. addend *= sign; int32_t x = (psymval->value(object, addend) | thumb_bit) - (address & 0xfffffffc); Reltype val = abs(x); // Mask out the value and a distinct part of the ADD/SUB opcode // (bits 7:5 of opword). insn = (insn & 0xfb0f8f00) | (val & 0xff) | ((val & 0x700) << 4) | ((val & 0x800) << 15); // Set the opcode according to whether the value to go in the // place is negative. if (x < 0) insn |= 0x00a00000; elfcpp::Swap<16, big_endian>::writeval(wv, insn >> 16); elfcpp::Swap<16, big_endian>::writeval(wv + 1, insn & 0xffff); return ((val > 0xfff) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_PC8: S + A - Pa (Thumb) static inline typename This::Status thm_pc8(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<16, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Valtype insn = elfcpp::Swap<16, big_endian>::readval(wv); Reltype addend = ((insn & 0x00ff) << 2); int32_t x = (psymval->value(object, addend) - (address & 0xfffffffc)); Reltype val = abs(x); insn = (insn & 0xff00) | ((val & 0x03fc) >> 2); elfcpp::Swap<16, big_endian>::writeval(wv, insn); return ((val > 0x03fc) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_THM_PC12: S + A - Pa (Thumb32) static inline typename This::Status thm_pc12(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, Arm_address address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; typedef typename elfcpp::Swap<32, big_endian>::Valtype Reltype; Valtype* wv = reinterpret_cast(view); Reltype insn = (elfcpp::Swap<16, big_endian>::readval(wv) << 16) | elfcpp::Swap<16, big_endian>::readval(wv + 1); // Determine a sign for the addend (positive if the U bit is 1). const int sign = (insn & 0x00800000) ? 1 : -1; int32_t addend = (insn & 0xfff); // Apply a sign to the added. addend *= sign; int32_t x = (psymval->value(object, addend) - (address & 0xfffffffc)); Reltype val = abs(x); // Mask out and apply the value and the U bit. insn = (insn & 0xff7ff000) | (val & 0xfff); // Set the U bit according to whether the value to go in the // place is positive. if (x >= 0) insn |= 0x00800000; elfcpp::Swap<16, big_endian>::writeval(wv, insn >> 16); elfcpp::Swap<16, big_endian>::writeval(wv + 1, insn & 0xffff); return ((val > 0xfff) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // R_ARM_V4BX static inline typename This::Status v4bx(const Relocate_info<32, big_endian>* relinfo, unsigned char *view, const Arm_relobj* object, const Arm_address address, const bool is_interworking) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); // Ensure that we have a BX instruction. gold_assert((val & 0x0ffffff0) == 0x012fff10); const uint32_t reg = (val & 0xf); if (is_interworking && reg != 0xf) { Stub_table* stub_table = object->stub_table(relinfo->data_shndx); gold_assert(stub_table != NULL); Arm_v4bx_stub* stub = stub_table->find_arm_v4bx_stub(reg); gold_assert(stub != NULL); int32_t veneer_address = stub_table->address() + stub->offset() - 8 - address; gold_assert((veneer_address <= ARM_MAX_FWD_BRANCH_OFFSET) && (veneer_address >= ARM_MAX_BWD_BRANCH_OFFSET)); // Replace with a branch to veneer (B ) val = (val & 0xf0000000) | 0x0a000000 | ((veneer_address >> 2) & 0x00ffffff); } else { // Preserve Rm (lowest four bits) and the condition code // (highest four bits). Other bits encode MOV PC,Rm. val = (val & 0xf000000f) | 0x01a0f000; } elfcpp::Swap<32, big_endian>::writeval(wv, val); return This::STATUS_OKAY; } // R_ARM_ALU_PC_G0_NC: ((S + A) | T) - P // R_ARM_ALU_PC_G0: ((S + A) | T) - P // R_ARM_ALU_PC_G1_NC: ((S + A) | T) - P // R_ARM_ALU_PC_G1: ((S + A) | T) - P // R_ARM_ALU_PC_G2: ((S + A) | T) - P // R_ARM_ALU_SB_G0_NC: ((S + A) | T) - B(S) // R_ARM_ALU_SB_G0: ((S + A) | T) - B(S) // R_ARM_ALU_SB_G1_NC: ((S + A) | T) - B(S) // R_ARM_ALU_SB_G1: ((S + A) | T) - B(S) // R_ARM_ALU_SB_G2: ((S + A) | T) - B(S) static inline typename This::Status arm_grp_alu(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, const int group, Arm_address address, Arm_address thumb_bit, bool check_overflow) { gold_assert(group >= 0 && group < 3); typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype insn = elfcpp::Swap<32, big_endian>::readval(wv); // ALU group relocations are allowed only for the ADD/SUB instructions. // (0x00800000 - ADD, 0x00400000 - SUB) const Valtype opcode = insn & 0x01e00000; if (opcode != 0x00800000 && opcode != 0x00400000) return This::STATUS_BAD_RELOC; // Determine a sign for the addend. const int sign = (opcode == 0x00800000) ? 1 : -1; // shifter = rotate_imm * 2 const uint32_t shifter = (insn & 0xf00) >> 7; // Initial addend value. int32_t addend = insn & 0xff; // Rotate addend right by shifter. addend = (addend >> shifter) | (addend << (32 - shifter)); // Apply a sign to the added. addend *= sign; int32_t x = ((psymval->value(object, addend) | thumb_bit) - address); Valtype gn = Arm_relocate_functions::calc_grp_gn(abs(x), group); // Check for overflow if required if (check_overflow && (Arm_relocate_functions::calc_grp_residual(abs(x), group) != 0)) return This::STATUS_OVERFLOW; // Mask out the value and the ADD/SUB part of the opcode; take care // not to destroy the S bit. insn &= 0xff1ff000; // Set the opcode according to whether the value to go in the // place is negative. insn |= ((x < 0) ? 0x00400000 : 0x00800000); // Encode the offset (encoded Gn). insn |= gn; elfcpp::Swap<32, big_endian>::writeval(wv, insn); return This::STATUS_OKAY; } // R_ARM_LDR_PC_G0: S + A - P // R_ARM_LDR_PC_G1: S + A - P // R_ARM_LDR_PC_G2: S + A - P // R_ARM_LDR_SB_G0: S + A - B(S) // R_ARM_LDR_SB_G1: S + A - B(S) // R_ARM_LDR_SB_G2: S + A - B(S) static inline typename This::Status arm_grp_ldr(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, const int group, Arm_address address) { gold_assert(group >= 0 && group < 3); typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype insn = elfcpp::Swap<32, big_endian>::readval(wv); const int sign = (insn & 0x00800000) ? 1 : -1; int32_t addend = (insn & 0xfff) * sign; int32_t x = (psymval->value(object, addend) - address); // Calculate the relevant G(n-1) value to obtain this stage residual. Valtype residual = Arm_relocate_functions::calc_grp_residual(abs(x), group - 1); if (residual >= 0x1000) return This::STATUS_OVERFLOW; // Mask out the value and U bit. insn &= 0xff7ff000; // Set the U bit for non-negative values. if (x >= 0) insn |= 0x00800000; insn |= residual; elfcpp::Swap<32, big_endian>::writeval(wv, insn); return This::STATUS_OKAY; } // R_ARM_LDRS_PC_G0: S + A - P // R_ARM_LDRS_PC_G1: S + A - P // R_ARM_LDRS_PC_G2: S + A - P // R_ARM_LDRS_SB_G0: S + A - B(S) // R_ARM_LDRS_SB_G1: S + A - B(S) // R_ARM_LDRS_SB_G2: S + A - B(S) static inline typename This::Status arm_grp_ldrs(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, const int group, Arm_address address) { gold_assert(group >= 0 && group < 3); typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype insn = elfcpp::Swap<32, big_endian>::readval(wv); const int sign = (insn & 0x00800000) ? 1 : -1; int32_t addend = (((insn & 0xf00) >> 4) + (insn & 0xf)) * sign; int32_t x = (psymval->value(object, addend) - address); // Calculate the relevant G(n-1) value to obtain this stage residual. Valtype residual = Arm_relocate_functions::calc_grp_residual(abs(x), group - 1); if (residual >= 0x100) return This::STATUS_OVERFLOW; // Mask out the value and U bit. insn &= 0xff7ff0f0; // Set the U bit for non-negative values. if (x >= 0) insn |= 0x00800000; insn |= ((residual & 0xf0) << 4) | (residual & 0xf); elfcpp::Swap<32, big_endian>::writeval(wv, insn); return This::STATUS_OKAY; } // R_ARM_LDC_PC_G0: S + A - P // R_ARM_LDC_PC_G1: S + A - P // R_ARM_LDC_PC_G2: S + A - P // R_ARM_LDC_SB_G0: S + A - B(S) // R_ARM_LDC_SB_G1: S + A - B(S) // R_ARM_LDC_SB_G2: S + A - B(S) static inline typename This::Status arm_grp_ldc(unsigned char* view, const Sized_relobj<32, big_endian>* object, const Symbol_value<32>* psymval, const int group, Arm_address address) { gold_assert(group >= 0 && group < 3); typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype insn = elfcpp::Swap<32, big_endian>::readval(wv); const int sign = (insn & 0x00800000) ? 1 : -1; int32_t addend = ((insn & 0xff) << 2) * sign; int32_t x = (psymval->value(object, addend) - address); // Calculate the relevant G(n-1) value to obtain this stage residual. Valtype residual = Arm_relocate_functions::calc_grp_residual(abs(x), group - 1); if ((residual & 0x3) != 0 || residual >= 0x400) return This::STATUS_OVERFLOW; // Mask out the value and U bit. insn &= 0xff7fff00; // Set the U bit for non-negative values. if (x >= 0) insn |= 0x00800000; insn |= (residual >> 2); elfcpp::Swap<32, big_endian>::writeval(wv, insn); return This::STATUS_OKAY; } }; // Relocate ARM long branches. This handles relocation types // R_ARM_CALL, R_ARM_JUMP24, R_ARM_PLT32 and R_ARM_XPC25. // If IS_WEAK_UNDEFINED_WITH_PLT is true. The target symbol is weakly // undefined and we do not use PLT in this relocation. In such a case, // the branch is converted into an NOP. template typename Arm_relocate_functions::Status Arm_relocate_functions::arm_branch_common( unsigned int r_type, const Relocate_info<32, big_endian>* relinfo, unsigned char *view, const Sized_symbol<32>* gsym, const Arm_relobj* object, unsigned int r_sym, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit, bool is_weakly_undefined_without_plt) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); bool insn_is_b = (((val >> 28) & 0xf) <= 0xe) && ((val & 0x0f000000UL) == 0x0a000000UL); bool insn_is_uncond_bl = (val & 0xff000000UL) == 0xeb000000UL; bool insn_is_cond_bl = (((val >> 28) & 0xf) < 0xe) && ((val & 0x0f000000UL) == 0x0b000000UL); bool insn_is_blx = (val & 0xfe000000UL) == 0xfa000000UL; bool insn_is_any_branch = (val & 0x0e000000UL) == 0x0a000000UL; // Check that the instruction is valid. if (r_type == elfcpp::R_ARM_CALL) { if (!insn_is_uncond_bl && !insn_is_blx) return This::STATUS_BAD_RELOC; } else if (r_type == elfcpp::R_ARM_JUMP24) { if (!insn_is_b && !insn_is_cond_bl) return This::STATUS_BAD_RELOC; } else if (r_type == elfcpp::R_ARM_PLT32) { if (!insn_is_any_branch) return This::STATUS_BAD_RELOC; } else if (r_type == elfcpp::R_ARM_XPC25) { // FIXME: AAELF document IH0044C does not say much about it other // than it being obsolete. if (!insn_is_any_branch) return This::STATUS_BAD_RELOC; } else gold_unreachable(); // A branch to an undefined weak symbol is turned into a jump to // the next instruction unless a PLT entry will be created. // Do the same for local undefined symbols. // The jump to the next instruction is optimized as a NOP depending // on the architecture. const Target_arm* arm_target = Target_arm::default_target(); if (is_weakly_undefined_without_plt) { Valtype cond = val & 0xf0000000U; if (arm_target->may_use_arm_nop()) val = cond | 0x0320f000; else val = cond | 0x01a00000; // Using pre-UAL nop: mov r0, r0. elfcpp::Swap<32, big_endian>::writeval(wv, val); return This::STATUS_OKAY; } Valtype addend = utils::sign_extend<26>(val << 2); Valtype branch_target = psymval->value(object, addend); int32_t branch_offset = branch_target - address; // We need a stub if the branch offset is too large or if we need // to switch mode. bool may_use_blx = arm_target->may_use_blx(); Reloc_stub* stub = NULL; if ((branch_offset > ARM_MAX_FWD_BRANCH_OFFSET) || (branch_offset < ARM_MAX_BWD_BRANCH_OFFSET) || ((thumb_bit != 0) && !(may_use_blx && r_type == elfcpp::R_ARM_CALL))) { Stub_type stub_type = Reloc_stub::stub_type_for_reloc(r_type, address, branch_target, (thumb_bit != 0)); if (stub_type != arm_stub_none) { Stub_table* stub_table = object->stub_table(relinfo->data_shndx); gold_assert(stub_table != NULL); Reloc_stub::Key stub_key(stub_type, gsym, object, r_sym, addend); stub = stub_table->find_reloc_stub(stub_key); gold_assert(stub != NULL); thumb_bit = stub->stub_template()->entry_in_thumb_mode() ? 1 : 0; branch_target = stub_table->address() + stub->offset() + addend; branch_offset = branch_target - address; gold_assert((branch_offset <= ARM_MAX_FWD_BRANCH_OFFSET) && (branch_offset >= ARM_MAX_BWD_BRANCH_OFFSET)); } } // At this point, if we still need to switch mode, the instruction // must either be a BLX or a BL that can be converted to a BLX. if (thumb_bit != 0) { // Turn BL to BLX. gold_assert(may_use_blx && r_type == elfcpp::R_ARM_CALL); val = (val & 0xffffff) | 0xfa000000 | ((branch_offset & 2) << 23); } val = utils::bit_select(val, (branch_offset >> 2), 0xffffffUL); elfcpp::Swap<32, big_endian>::writeval(wv, val); return (utils::has_overflow<26>(branch_offset) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // Relocate THUMB long branches. This handles relocation types // R_ARM_THM_CALL, R_ARM_THM_JUMP24 and R_ARM_THM_XPC22. // If IS_WEAK_UNDEFINED_WITH_PLT is true. The target symbol is weakly // undefined and we do not use PLT in this relocation. In such a case, // the branch is converted into an NOP. template typename Arm_relocate_functions::Status Arm_relocate_functions::thumb_branch_common( unsigned int r_type, const Relocate_info<32, big_endian>* relinfo, unsigned char *view, const Sized_symbol<32>* gsym, const Arm_relobj* object, unsigned int r_sym, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit, bool is_weakly_undefined_without_plt) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); uint32_t upper_insn = elfcpp::Swap<16, big_endian>::readval(wv); uint32_t lower_insn = elfcpp::Swap<16, big_endian>::readval(wv + 1); // FIXME: These tests are too loose and do not take THUMB/THUMB-2 difference // into account. bool is_bl_insn = (lower_insn & 0x1000U) == 0x1000U; bool is_blx_insn = (lower_insn & 0x1000U) == 0x0000U; // Check that the instruction is valid. if (r_type == elfcpp::R_ARM_THM_CALL) { if (!is_bl_insn && !is_blx_insn) return This::STATUS_BAD_RELOC; } else if (r_type == elfcpp::R_ARM_THM_JUMP24) { // This cannot be a BLX. if (!is_bl_insn) return This::STATUS_BAD_RELOC; } else if (r_type == elfcpp::R_ARM_THM_XPC22) { // Check for Thumb to Thumb call. if (!is_blx_insn) return This::STATUS_BAD_RELOC; if (thumb_bit != 0) { gold_warning(_("%s: Thumb BLX instruction targets " "thumb function '%s'."), object->name().c_str(), (gsym ? gsym->name() : "(local)")); // Convert BLX to BL. lower_insn |= 0x1000U; } } else gold_unreachable(); // A branch to an undefined weak symbol is turned into a jump to // the next instruction unless a PLT entry will be created. // The jump to the next instruction is optimized as a NOP.W for // Thumb-2 enabled architectures. const Target_arm* arm_target = Target_arm::default_target(); if (is_weakly_undefined_without_plt) { if (arm_target->may_use_thumb2_nop()) { elfcpp::Swap<16, big_endian>::writeval(wv, 0xf3af); elfcpp::Swap<16, big_endian>::writeval(wv + 1, 0x8000); } else { elfcpp::Swap<16, big_endian>::writeval(wv, 0xe000); elfcpp::Swap<16, big_endian>::writeval(wv + 1, 0xbf00); } return This::STATUS_OKAY; } int32_t addend = This::thumb32_branch_offset(upper_insn, lower_insn); Arm_address branch_target = psymval->value(object, addend); int32_t branch_offset = branch_target - address; // We need a stub if the branch offset is too large or if we need // to switch mode. bool may_use_blx = arm_target->may_use_blx(); bool thumb2 = arm_target->using_thumb2(); if ((!thumb2 && (branch_offset > THM_MAX_FWD_BRANCH_OFFSET || (branch_offset < THM_MAX_BWD_BRANCH_OFFSET))) || (thumb2 && (branch_offset > THM2_MAX_FWD_BRANCH_OFFSET || (branch_offset < THM2_MAX_BWD_BRANCH_OFFSET))) || ((thumb_bit == 0) && (((r_type == elfcpp::R_ARM_THM_CALL) && !may_use_blx) || r_type == elfcpp::R_ARM_THM_JUMP24))) { Stub_type stub_type = Reloc_stub::stub_type_for_reloc(r_type, address, branch_target, (thumb_bit != 0)); if (stub_type != arm_stub_none) { Stub_table* stub_table = object->stub_table(relinfo->data_shndx); gold_assert(stub_table != NULL); Reloc_stub::Key stub_key(stub_type, gsym, object, r_sym, addend); Reloc_stub* stub = stub_table->find_reloc_stub(stub_key); gold_assert(stub != NULL); thumb_bit = stub->stub_template()->entry_in_thumb_mode() ? 1 : 0; branch_target = stub_table->address() + stub->offset() + addend; branch_offset = branch_target - address; } } // At this point, if we still need to switch mode, the instruction // must either be a BLX or a BL that can be converted to a BLX. if (thumb_bit == 0) { gold_assert(may_use_blx && (r_type == elfcpp::R_ARM_THM_CALL || r_type == elfcpp::R_ARM_THM_XPC22)); // Make sure this is a BLX. lower_insn &= ~0x1000U; } else { // Make sure this is a BL. lower_insn |= 0x1000U; } if ((lower_insn & 0x5000U) == 0x4000U) // For a BLX instruction, make sure that the relocation is rounded up // to a word boundary. This follows the semantics of the instruction // which specifies that bit 1 of the target address will come from bit // 1 of the base address. branch_offset = (branch_offset + 2) & ~3; // Put BRANCH_OFFSET back into the insn. Assumes two's complement. // We use the Thumb-2 encoding, which is safe even if dealing with // a Thumb-1 instruction by virtue of our overflow check above. */ upper_insn = This::thumb32_branch_upper(upper_insn, branch_offset); lower_insn = This::thumb32_branch_lower(lower_insn, branch_offset); elfcpp::Swap<16, big_endian>::writeval(wv, upper_insn); elfcpp::Swap<16, big_endian>::writeval(wv + 1, lower_insn); return ((thumb2 ? utils::has_overflow<25>(branch_offset) : utils::has_overflow<23>(branch_offset)) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // Relocate THUMB-2 long conditional branches. // If IS_WEAK_UNDEFINED_WITH_PLT is true. The target symbol is weakly // undefined and we do not use PLT in this relocation. In such a case, // the branch is converted into an NOP. template typename Arm_relocate_functions::Status Arm_relocate_functions::thm_jump19( unsigned char *view, const Arm_relobj* object, const Symbol_value<32>* psymval, Arm_address address, Arm_address thumb_bit) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(view); uint32_t upper_insn = elfcpp::Swap<16, big_endian>::readval(wv); uint32_t lower_insn = elfcpp::Swap<16, big_endian>::readval(wv + 1); int32_t addend = This::thumb32_cond_branch_offset(upper_insn, lower_insn); Arm_address branch_target = psymval->value(object, addend); int32_t branch_offset = branch_target - address; // ??? Should handle interworking? GCC might someday try to // use this for tail calls. // FIXME: We do support thumb entry to PLT yet. if (thumb_bit == 0) { gold_error(_("conditional branch to PLT in THUMB-2 not supported yet.")); return This::STATUS_BAD_RELOC; } // Put RELOCATION back into the insn. upper_insn = This::thumb32_cond_branch_upper(upper_insn, branch_offset); lower_insn = This::thumb32_cond_branch_lower(lower_insn, branch_offset); // Put the relocated value back in the object file: elfcpp::Swap<16, big_endian>::writeval(wv, upper_insn); elfcpp::Swap<16, big_endian>::writeval(wv + 1, lower_insn); return (utils::has_overflow<21>(branch_offset) ? This::STATUS_OVERFLOW : This::STATUS_OKAY); } // Get the GOT section, creating it if necessary. template Output_data_got<32, big_endian>* Target_arm::got_section(Symbol_table* symtab, Layout* layout) { if (this->got_ == NULL) { gold_assert(symtab != NULL && layout != NULL); this->got_ = new Output_data_got<32, big_endian>(); Output_section* os; os = layout->add_output_section_data(".got", elfcpp::SHT_PROGBITS, (elfcpp::SHF_ALLOC | elfcpp::SHF_WRITE), this->got_, false, true, true, false); // The old GNU linker creates a .got.plt section. We just // create another set of data in the .got section. Note that we // always create a PLT if we create a GOT, although the PLT // might be empty. this->got_plt_ = new Output_data_space(4, "** GOT PLT"); os = layout->add_output_section_data(".got", elfcpp::SHT_PROGBITS, (elfcpp::SHF_ALLOC | elfcpp::SHF_WRITE), this->got_plt_, false, false, false, true); // The first three entries are reserved. this->got_plt_->set_current_data_size(3 * 4); // Define _GLOBAL_OFFSET_TABLE_ at the start of the PLT. symtab->define_in_output_data("_GLOBAL_OFFSET_TABLE_", NULL, Symbol_table::PREDEFINED, this->got_plt_, 0, 0, elfcpp::STT_OBJECT, elfcpp::STB_LOCAL, elfcpp::STV_HIDDEN, 0, false, false); } return this->got_; } // Get the dynamic reloc section, creating it if necessary. template typename Target_arm::Reloc_section* Target_arm::rel_dyn_section(Layout* layout) { if (this->rel_dyn_ == NULL) { gold_assert(layout != NULL); this->rel_dyn_ = new Reloc_section(parameters->options().combreloc()); layout->add_output_section_data(".rel.dyn", elfcpp::SHT_REL, elfcpp::SHF_ALLOC, this->rel_dyn_, true, false, false, false); } return this->rel_dyn_; } // Insn_template methods. // Return byte size of an instruction template. size_t Insn_template::size() const { switch (this->type()) { case THUMB16_TYPE: case THUMB16_SPECIAL_TYPE: return 2; case ARM_TYPE: case THUMB32_TYPE: case DATA_TYPE: return 4; default: gold_unreachable(); } } // Return alignment of an instruction template. unsigned Insn_template::alignment() const { switch (this->type()) { case THUMB16_TYPE: case THUMB16_SPECIAL_TYPE: case THUMB32_TYPE: return 2; case ARM_TYPE: case DATA_TYPE: return 4; default: gold_unreachable(); } } // Stub_template methods. Stub_template::Stub_template( Stub_type type, const Insn_template* insns, size_t insn_count) : type_(type), insns_(insns), insn_count_(insn_count), alignment_(1), entry_in_thumb_mode_(false), relocs_() { off_t offset = 0; // Compute byte size and alignment of stub template. for (size_t i = 0; i < insn_count; i++) { unsigned insn_alignment = insns[i].alignment(); size_t insn_size = insns[i].size(); gold_assert((offset & (insn_alignment - 1)) == 0); this->alignment_ = std::max(this->alignment_, insn_alignment); switch (insns[i].type()) { case Insn_template::THUMB16_TYPE: case Insn_template::THUMB16_SPECIAL_TYPE: if (i == 0) this->entry_in_thumb_mode_ = true; break; case Insn_template::THUMB32_TYPE: if (insns[i].r_type() != elfcpp::R_ARM_NONE) this->relocs_.push_back(Reloc(i, offset)); if (i == 0) this->entry_in_thumb_mode_ = true; break; case Insn_template::ARM_TYPE: // Handle cases where the target is encoded within the // instruction. if (insns[i].r_type() == elfcpp::R_ARM_JUMP24) this->relocs_.push_back(Reloc(i, offset)); break; case Insn_template::DATA_TYPE: // Entry point cannot be data. gold_assert(i != 0); this->relocs_.push_back(Reloc(i, offset)); break; default: gold_unreachable(); } offset += insn_size; } this->size_ = offset; } // Stub methods. // Template to implement do_write for a specific target endianity. template void inline Stub::do_fixed_endian_write(unsigned char* view, section_size_type view_size) { const Stub_template* stub_template = this->stub_template(); const Insn_template* insns = stub_template->insns(); // FIXME: We do not handle BE8 encoding yet. unsigned char* pov = view; for (size_t i = 0; i < stub_template->insn_count(); i++) { switch (insns[i].type()) { case Insn_template::THUMB16_TYPE: elfcpp::Swap<16, big_endian>::writeval(pov, insns[i].data() & 0xffff); break; case Insn_template::THUMB16_SPECIAL_TYPE: elfcpp::Swap<16, big_endian>::writeval( pov, this->thumb16_special(i)); break; case Insn_template::THUMB32_TYPE: { uint32_t hi = (insns[i].data() >> 16) & 0xffff; uint32_t lo = insns[i].data() & 0xffff; elfcpp::Swap<16, big_endian>::writeval(pov, hi); elfcpp::Swap<16, big_endian>::writeval(pov + 2, lo); } break; case Insn_template::ARM_TYPE: case Insn_template::DATA_TYPE: elfcpp::Swap<32, big_endian>::writeval(pov, insns[i].data()); break; default: gold_unreachable(); } pov += insns[i].size(); } gold_assert(static_cast(pov - view) == view_size); } // Reloc_stub::Key methods. // Dump a Key as a string for debugging. std::string Reloc_stub::Key::name() const { if (this->r_sym_ == invalid_index) { // Global symbol key name // ::. const std::string sym_name = this->u_.symbol->name(); // We need to print two hex number and two colons. So just add 100 bytes // to the symbol name size. size_t len = sym_name.size() + 100; char* buffer = new char[len]; int c = snprintf(buffer, len, "%d:%s:%x", this->stub_type_, sym_name.c_str(), this->addend_); gold_assert(c > 0 && c < static_cast(len)); delete[] buffer; return std::string(buffer); } else { // local symbol key name // :::. const size_t len = 200; char buffer[len]; int c = snprintf(buffer, len, "%d:%p:%u:%x", this->stub_type_, this->u_.relobj, this->r_sym_, this->addend_); gold_assert(c > 0 && c < static_cast(len)); return std::string(buffer); } } // Reloc_stub methods. // Determine the type of stub needed, if any, for a relocation of R_TYPE at // LOCATION to DESTINATION. // This code is based on the arm_type_of_stub function in // bfd/elf32-arm.c. We have changed the interface a liitle to keep the Stub // class simple. Stub_type Reloc_stub::stub_type_for_reloc( unsigned int r_type, Arm_address location, Arm_address destination, bool target_is_thumb) { Stub_type stub_type = arm_stub_none; // This is a bit ugly but we want to avoid using a templated class for // big and little endianities. bool may_use_blx; bool should_force_pic_veneer; bool thumb2; bool thumb_only; if (parameters->target().is_big_endian()) { const Target_arm* big_endian_target = Target_arm::default_target(); may_use_blx = big_endian_target->may_use_blx(); should_force_pic_veneer = big_endian_target->should_force_pic_veneer(); thumb2 = big_endian_target->using_thumb2(); thumb_only = big_endian_target->using_thumb_only(); } else { const Target_arm* little_endian_target = Target_arm::default_target(); may_use_blx = little_endian_target->may_use_blx(); should_force_pic_veneer = little_endian_target->should_force_pic_veneer(); thumb2 = little_endian_target->using_thumb2(); thumb_only = little_endian_target->using_thumb_only(); } int64_t branch_offset = (int64_t)destination - location; if (r_type == elfcpp::R_ARM_THM_CALL || r_type == elfcpp::R_ARM_THM_JUMP24) { // Handle cases where: // - this call goes too far (different Thumb/Thumb2 max // distance) // - it's a Thumb->Arm call and blx is not available, or it's a // Thumb->Arm branch (not bl). A stub is needed in this case. if ((!thumb2 && (branch_offset > THM_MAX_FWD_BRANCH_OFFSET || (branch_offset < THM_MAX_BWD_BRANCH_OFFSET))) || (thumb2 && (branch_offset > THM2_MAX_FWD_BRANCH_OFFSET || (branch_offset < THM2_MAX_BWD_BRANCH_OFFSET))) || ((!target_is_thumb) && (((r_type == elfcpp::R_ARM_THM_CALL) && !may_use_blx) || (r_type == elfcpp::R_ARM_THM_JUMP24)))) { if (target_is_thumb) { // Thumb to thumb. if (!thumb_only) { stub_type = (parameters->options().shared() || should_force_pic_veneer) // PIC stubs. ? ((may_use_blx && (r_type == elfcpp::R_ARM_THM_CALL)) // V5T and above. Stub starts with ARM code, so // we must be able to switch mode before // reaching it, which is only possible for 'bl' // (ie R_ARM_THM_CALL relocation). ? arm_stub_long_branch_any_thumb_pic // On V4T, use Thumb code only. : arm_stub_long_branch_v4t_thumb_thumb_pic) // non-PIC stubs. : ((may_use_blx && (r_type == elfcpp::R_ARM_THM_CALL)) ? arm_stub_long_branch_any_any // V5T and above. : arm_stub_long_branch_v4t_thumb_thumb); // V4T. } else { stub_type = (parameters->options().shared() || should_force_pic_veneer) ? arm_stub_long_branch_thumb_only_pic // PIC stub. : arm_stub_long_branch_thumb_only; // non-PIC stub. } } else { // Thumb to arm. // FIXME: We should check that the input section is from an // object that has interwork enabled. stub_type = (parameters->options().shared() || should_force_pic_veneer) // PIC stubs. ? ((may_use_blx && (r_type == elfcpp::R_ARM_THM_CALL)) ? arm_stub_long_branch_any_arm_pic // V5T and above. : arm_stub_long_branch_v4t_thumb_arm_pic) // V4T. // non-PIC stubs. : ((may_use_blx && (r_type == elfcpp::R_ARM_THM_CALL)) ? arm_stub_long_branch_any_any // V5T and above. : arm_stub_long_branch_v4t_thumb_arm); // V4T. // Handle v4t short branches. if ((stub_type == arm_stub_long_branch_v4t_thumb_arm) && (branch_offset <= THM_MAX_FWD_BRANCH_OFFSET) && (branch_offset >= THM_MAX_BWD_BRANCH_OFFSET)) stub_type = arm_stub_short_branch_v4t_thumb_arm; } } } else if (r_type == elfcpp::R_ARM_CALL || r_type == elfcpp::R_ARM_JUMP24 || r_type == elfcpp::R_ARM_PLT32) { if (target_is_thumb) { // Arm to thumb. // FIXME: We should check that the input section is from an // object that has interwork enabled. // We have an extra 2-bytes reach because of // the mode change (bit 24 (H) of BLX encoding). if (branch_offset > (ARM_MAX_FWD_BRANCH_OFFSET + 2) || (branch_offset < ARM_MAX_BWD_BRANCH_OFFSET) || ((r_type == elfcpp::R_ARM_CALL) && !may_use_blx) || (r_type == elfcpp::R_ARM_JUMP24) || (r_type == elfcpp::R_ARM_PLT32)) { stub_type = (parameters->options().shared() || should_force_pic_veneer) // PIC stubs. ? (may_use_blx ? arm_stub_long_branch_any_thumb_pic// V5T and above. : arm_stub_long_branch_v4t_arm_thumb_pic) // V4T stub. // non-PIC stubs. : (may_use_blx ? arm_stub_long_branch_any_any // V5T and above. : arm_stub_long_branch_v4t_arm_thumb); // V4T. } } else { // Arm to arm. if (branch_offset > ARM_MAX_FWD_BRANCH_OFFSET || (branch_offset < ARM_MAX_BWD_BRANCH_OFFSET)) { stub_type = (parameters->options().shared() || should_force_pic_veneer) ? arm_stub_long_branch_any_arm_pic // PIC stubs. : arm_stub_long_branch_any_any; /// non-PIC. } } } return stub_type; } // Cortex_a8_stub methods. // Return the instruction for a THUMB16_SPECIAL_TYPE instruction template. // I is the position of the instruction template in the stub template. uint16_t Cortex_a8_stub::do_thumb16_special(size_t i) { // The only use of this is to copy condition code from a conditional // branch being worked around to the corresponding conditional branch in // to the stub. gold_assert(this->stub_template()->type() == arm_stub_a8_veneer_b_cond && i == 0); uint16_t data = this->stub_template()->insns()[i].data(); gold_assert((data & 0xff00U) == 0xd000U); data |= ((this->original_insn_ >> 22) & 0xf) << 8; return data; } // Stub_factory methods. Stub_factory::Stub_factory() { // The instruction template sequences are declared as static // objects and initialized first time the constructor runs. // Arm/Thumb -> Arm/Thumb long branch stub. On V5T and above, use blx // to reach the stub if necessary. static const Insn_template elf32_arm_stub_long_branch_any_any[] = { Insn_template::arm_insn(0xe51ff004), // ldr pc, [pc, #-4] Insn_template::data_word(0, elfcpp::R_ARM_ABS32, 0), // dcd R_ARM_ABS32(X) }; // V4T Arm -> Thumb long branch stub. Used on V4T where blx is not // available. static const Insn_template elf32_arm_stub_long_branch_v4t_arm_thumb[] = { Insn_template::arm_insn(0xe59fc000), // ldr ip, [pc, #0] Insn_template::arm_insn(0xe12fff1c), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_ABS32, 0), // dcd R_ARM_ABS32(X) }; // Thumb -> Thumb long branch stub. Used on M-profile architectures. static const Insn_template elf32_arm_stub_long_branch_thumb_only[] = { Insn_template::thumb16_insn(0xb401), // push {r0} Insn_template::thumb16_insn(0x4802), // ldr r0, [pc, #8] Insn_template::thumb16_insn(0x4684), // mov ip, r0 Insn_template::thumb16_insn(0xbc01), // pop {r0} Insn_template::thumb16_insn(0x4760), // bx ip Insn_template::thumb16_insn(0xbf00), // nop Insn_template::data_word(0, elfcpp::R_ARM_ABS32, 0), // dcd R_ARM_ABS32(X) }; // V4T Thumb -> Thumb long branch stub. Using the stack is not // allowed. static const Insn_template elf32_arm_stub_long_branch_v4t_thumb_thumb[] = { Insn_template::thumb16_insn(0x4778), // bx pc Insn_template::thumb16_insn(0x46c0), // nop Insn_template::arm_insn(0xe59fc000), // ldr ip, [pc, #0] Insn_template::arm_insn(0xe12fff1c), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_ABS32, 0), // dcd R_ARM_ABS32(X) }; // V4T Thumb -> ARM long branch stub. Used on V4T where blx is not // available. static const Insn_template elf32_arm_stub_long_branch_v4t_thumb_arm[] = { Insn_template::thumb16_insn(0x4778), // bx pc Insn_template::thumb16_insn(0x46c0), // nop Insn_template::arm_insn(0xe51ff004), // ldr pc, [pc, #-4] Insn_template::data_word(0, elfcpp::R_ARM_ABS32, 0), // dcd R_ARM_ABS32(X) }; // V4T Thumb -> ARM short branch stub. Shorter variant of the above // one, when the destination is close enough. static const Insn_template elf32_arm_stub_short_branch_v4t_thumb_arm[] = { Insn_template::thumb16_insn(0x4778), // bx pc Insn_template::thumb16_insn(0x46c0), // nop Insn_template::arm_rel_insn(0xea000000, -8), // b (X-8) }; // ARM/Thumb -> ARM long branch stub, PIC. On V5T and above, use // blx to reach the stub if necessary. static const Insn_template elf32_arm_stub_long_branch_any_arm_pic[] = { Insn_template::arm_insn(0xe59fc000), // ldr r12, [pc] Insn_template::arm_insn(0xe08ff00c), // add pc, pc, ip Insn_template::data_word(0, elfcpp::R_ARM_REL32, -4), // dcd R_ARM_REL32(X-4) }; // ARM/Thumb -> Thumb long branch stub, PIC. On V5T and above, use // blx to reach the stub if necessary. We can not add into pc; // it is not guaranteed to mode switch (different in ARMv6 and // ARMv7). static const Insn_template elf32_arm_stub_long_branch_any_thumb_pic[] = { Insn_template::arm_insn(0xe59fc004), // ldr r12, [pc, #4] Insn_template::arm_insn(0xe08fc00c), // add ip, pc, ip Insn_template::arm_insn(0xe12fff1c), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_REL32, 0), // dcd R_ARM_REL32(X) }; // V4T ARM -> ARM long branch stub, PIC. static const Insn_template elf32_arm_stub_long_branch_v4t_arm_thumb_pic[] = { Insn_template::arm_insn(0xe59fc004), // ldr ip, [pc, #4] Insn_template::arm_insn(0xe08fc00c), // add ip, pc, ip Insn_template::arm_insn(0xe12fff1c), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_REL32, 0), // dcd R_ARM_REL32(X) }; // V4T Thumb -> ARM long branch stub, PIC. static const Insn_template elf32_arm_stub_long_branch_v4t_thumb_arm_pic[] = { Insn_template::thumb16_insn(0x4778), // bx pc Insn_template::thumb16_insn(0x46c0), // nop Insn_template::arm_insn(0xe59fc000), // ldr ip, [pc, #0] Insn_template::arm_insn(0xe08cf00f), // add pc, ip, pc Insn_template::data_word(0, elfcpp::R_ARM_REL32, -4), // dcd R_ARM_REL32(X) }; // Thumb -> Thumb long branch stub, PIC. Used on M-profile // architectures. static const Insn_template elf32_arm_stub_long_branch_thumb_only_pic[] = { Insn_template::thumb16_insn(0xb401), // push {r0} Insn_template::thumb16_insn(0x4802), // ldr r0, [pc, #8] Insn_template::thumb16_insn(0x46fc), // mov ip, pc Insn_template::thumb16_insn(0x4484), // add ip, r0 Insn_template::thumb16_insn(0xbc01), // pop {r0} Insn_template::thumb16_insn(0x4760), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_REL32, 4), // dcd R_ARM_REL32(X) }; // V4T Thumb -> Thumb long branch stub, PIC. Using the stack is not // allowed. static const Insn_template elf32_arm_stub_long_branch_v4t_thumb_thumb_pic[] = { Insn_template::thumb16_insn(0x4778), // bx pc Insn_template::thumb16_insn(0x46c0), // nop Insn_template::arm_insn(0xe59fc004), // ldr ip, [pc, #4] Insn_template::arm_insn(0xe08fc00c), // add ip, pc, ip Insn_template::arm_insn(0xe12fff1c), // bx ip Insn_template::data_word(0, elfcpp::R_ARM_REL32, 0), // dcd R_ARM_REL32(X) }; // Cortex-A8 erratum-workaround stubs. // Stub used for conditional branches (which may be beyond +/-1MB away, // so we can't use a conditional branch to reach this stub). // original code: // // b X // after: // static const Insn_template elf32_arm_stub_a8_veneer_b_cond[] = { Insn_template::thumb16_bcond_insn(0xd001), // b.n true Insn_template::thumb32_b_insn(0xf000b800, -4), // b.w after Insn_template::thumb32_b_insn(0xf000b800, -4) // true: // b.w X }; // Stub used for b.w and bl.w instructions. static const Insn_template elf32_arm_stub_a8_veneer_b[] = { Insn_template::thumb32_b_insn(0xf000b800, -4) // b.w dest }; static const Insn_template elf32_arm_stub_a8_veneer_bl[] = { Insn_template::thumb32_b_insn(0xf000b800, -4) // b.w dest }; // Stub used for Thumb-2 blx.w instructions. We modified the original blx.w // instruction (which switches to ARM mode) to point to this stub. Jump to // the real destination using an ARM-mode branch. static const Insn_template elf32_arm_stub_a8_veneer_blx[] = { Insn_template::arm_rel_insn(0xea000000, -8) // b dest }; // Stub used to provide an interworking for R_ARM_V4BX relocation // (bx r[n] instruction). static const Insn_template elf32_arm_stub_v4_veneer_bx[] = { Insn_template::arm_insn(0xe3100001), // tst r, #1 Insn_template::arm_insn(0x01a0f000), // moveq pc, r Insn_template::arm_insn(0xe12fff10) // bx r }; // Fill in the stub template look-up table. Stub templates are constructed // per instance of Stub_factory for fast look-up without locking // in a thread-enabled environment. this->stub_templates_[arm_stub_none] = new Stub_template(arm_stub_none, NULL, 0); #define DEF_STUB(x) \ do \ { \ size_t array_size \ = sizeof(elf32_arm_stub_##x) / sizeof(elf32_arm_stub_##x[0]); \ Stub_type type = arm_stub_##x; \ this->stub_templates_[type] = \ new Stub_template(type, elf32_arm_stub_##x, array_size); \ } \ while (0); DEF_STUBS #undef DEF_STUB } // Stub_table methods. // Removel all Cortex-A8 stub. template void Stub_table::remove_all_cortex_a8_stubs() { for (Cortex_a8_stub_list::iterator p = this->cortex_a8_stubs_.begin(); p != this->cortex_a8_stubs_.end(); ++p) delete p->second; this->cortex_a8_stubs_.clear(); } // Relocate one stub. This is a helper for Stub_table::relocate_stubs(). template void Stub_table::relocate_stub( Stub* stub, const Relocate_info<32, big_endian>* relinfo, Target_arm* arm_target, Output_section* output_section, unsigned char* view, Arm_address address, section_size_type view_size) { const Stub_template* stub_template = stub->stub_template(); if (stub_template->reloc_count() != 0) { // Adjust view to cover the stub only. section_size_type offset = stub->offset(); section_size_type stub_size = stub_template->size(); gold_assert(offset + stub_size <= view_size); arm_target->relocate_stub(stub, relinfo, output_section, view + offset, address + offset, stub_size); } } // Relocate all stubs in this stub table. template void Stub_table::relocate_stubs( const Relocate_info<32, big_endian>* relinfo, Target_arm* arm_target, Output_section* output_section, unsigned char* view, Arm_address address, section_size_type view_size) { // If we are passed a view bigger than the stub table's. we need to // adjust the view. gold_assert(address == this->address() && (view_size == static_cast(this->data_size()))); // Relocate all relocation stubs. for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin(); p != this->reloc_stubs_.end(); ++p) this->relocate_stub(p->second, relinfo, arm_target, output_section, view, address, view_size); // Relocate all Cortex-A8 stubs. for (Cortex_a8_stub_list::iterator p = this->cortex_a8_stubs_.begin(); p != this->cortex_a8_stubs_.end(); ++p) this->relocate_stub(p->second, relinfo, arm_target, output_section, view, address, view_size); // Relocate all ARM V4BX stubs. for (Arm_v4bx_stub_list::iterator p = this->arm_v4bx_stubs_.begin(); p != this->arm_v4bx_stubs_.end(); ++p) { if (*p != NULL) this->relocate_stub(*p, relinfo, arm_target, output_section, view, address, view_size); } } // Write out the stubs to file. template void Stub_table::do_write(Output_file* of) { off_t offset = this->offset(); const section_size_type oview_size = convert_to_section_size_type(this->data_size()); unsigned char* const oview = of->get_output_view(offset, oview_size); // Write relocation stubs. for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin(); p != this->reloc_stubs_.end(); ++p) { Reloc_stub* stub = p->second; Arm_address address = this->address() + stub->offset(); gold_assert(address == align_address(address, stub->stub_template()->alignment())); stub->write(oview + stub->offset(), stub->stub_template()->size(), big_endian); } // Write Cortex-A8 stubs. for (Cortex_a8_stub_list::const_iterator p = this->cortex_a8_stubs_.begin(); p != this->cortex_a8_stubs_.end(); ++p) { Cortex_a8_stub* stub = p->second; Arm_address address = this->address() + stub->offset(); gold_assert(address == align_address(address, stub->stub_template()->alignment())); stub->write(oview + stub->offset(), stub->stub_template()->size(), big_endian); } // Write ARM V4BX relocation stubs. for (Arm_v4bx_stub_list::const_iterator p = this->arm_v4bx_stubs_.begin(); p != this->arm_v4bx_stubs_.end(); ++p) { if (*p == NULL) continue; Arm_address address = this->address() + (*p)->offset(); gold_assert(address == align_address(address, (*p)->stub_template()->alignment())); (*p)->write(oview + (*p)->offset(), (*p)->stub_template()->size(), big_endian); } of->write_output_view(this->offset(), oview_size, oview); } // Update the data size and address alignment of the stub table at the end // of a relaxation pass. Return true if either the data size or the // alignment changed in this relaxation pass. template bool Stub_table::update_data_size_and_addralign() { off_t size = 0; unsigned addralign = 1; // Go over all stubs in table to compute data size and address alignment. for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin(); p != this->reloc_stubs_.end(); ++p) { const Stub_template* stub_template = p->second->stub_template(); addralign = std::max(addralign, stub_template->alignment()); size = (align_address(size, stub_template->alignment()) + stub_template->size()); } for (Cortex_a8_stub_list::const_iterator p = this->cortex_a8_stubs_.begin(); p != this->cortex_a8_stubs_.end(); ++p) { const Stub_template* stub_template = p->second->stub_template(); addralign = std::max(addralign, stub_template->alignment()); size = (align_address(size, stub_template->alignment()) + stub_template->size()); } for (Arm_v4bx_stub_list::const_iterator p = this->arm_v4bx_stubs_.begin(); p != this->arm_v4bx_stubs_.end(); ++p) { if (*p == NULL) continue; const Stub_template* stub_template = (*p)->stub_template(); addralign = std::max(addralign, stub_template->alignment()); size = (align_address(size, stub_template->alignment()) + stub_template->size()); } // Check if either data size or alignment changed in this pass. // Update prev_data_size_ and prev_addralign_. These will be used // as the current data size and address alignment for the next pass. bool changed = size != this->prev_data_size_; this->prev_data_size_ = size; if (addralign != this->prev_addralign_) changed = true; this->prev_addralign_ = addralign; return changed; } // Finalize the stubs. This sets the offsets of the stubs within the stub // table. It also marks all input sections needing Cortex-A8 workaround. template void Stub_table::finalize_stubs() { off_t off = 0; for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin(); p != this->reloc_stubs_.end(); ++p) { Reloc_stub* stub = p->second; const Stub_template* stub_template = stub->stub_template(); uint64_t stub_addralign = stub_template->alignment(); off = align_address(off, stub_addralign); stub->set_offset(off); off += stub_template->size(); } for (Cortex_a8_stub_list::const_iterator p = this->cortex_a8_stubs_.begin(); p != this->cortex_a8_stubs_.end(); ++p) { Cortex_a8_stub* stub = p->second; const Stub_template* stub_template = stub->stub_template(); uint64_t stub_addralign = stub_template->alignment(); off = align_address(off, stub_addralign); stub->set_offset(off); off += stub_template->size(); // Mark input section so that we can determine later if a code section // needs the Cortex-A8 workaround quickly. Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(stub->relobj()); arm_relobj->mark_section_for_cortex_a8_workaround(stub->shndx()); } for (Arm_v4bx_stub_list::const_iterator p = this->arm_v4bx_stubs_.begin(); p != this->arm_v4bx_stubs_.end(); ++p) { if (*p == NULL) continue; const Stub_template* stub_template = (*p)->stub_template(); uint64_t stub_addralign = stub_template->alignment(); off = align_address(off, stub_addralign); (*p)->set_offset(off); off += stub_template->size(); } gold_assert(off <= this->prev_data_size_); } // Apply Cortex-A8 workaround to an address range between VIEW_ADDRESS // and VIEW_ADDRESS + VIEW_SIZE - 1. VIEW points to the mapped address // of the address range seen by the linker. template void Stub_table::apply_cortex_a8_workaround_to_address_range( Target_arm* arm_target, unsigned char* view, Arm_address view_address, section_size_type view_size) { // Cortex-A8 stubs are sorted by addresses of branches being fixed up. for (Cortex_a8_stub_list::const_iterator p = this->cortex_a8_stubs_.lower_bound(view_address); ((p != this->cortex_a8_stubs_.end()) && (p->first < (view_address + view_size))); ++p) { // We do not store the THUMB bit in the LSB of either the branch address // or the stub offset. There is no need to strip the LSB. Arm_address branch_address = p->first; const Cortex_a8_stub* stub = p->second; Arm_address stub_address = this->address() + stub->offset(); // Offset of the branch instruction relative to this view. section_size_type offset = convert_to_section_size_type(branch_address - view_address); gold_assert((offset + 4) <= view_size); arm_target->apply_cortex_a8_workaround(stub, stub_address, view + offset, branch_address); } } // Arm_input_section methods. // Initialize an Arm_input_section. template void Arm_input_section::init() { Relobj* relobj = this->relobj(); unsigned int shndx = this->shndx(); // Cache these to speed up size and alignment queries. It is too slow // to call section_addraglin and section_size every time. this->original_addralign_ = relobj->section_addralign(shndx); this->original_size_ = relobj->section_size(shndx); // We want to make this look like the original input section after // output sections are finalized. Output_section* os = relobj->output_section(shndx); off_t offset = relobj->output_section_offset(shndx); gold_assert(os != NULL && !relobj->is_output_section_offset_invalid(shndx)); this->set_address(os->address() + offset); this->set_file_offset(os->offset() + offset); this->set_current_data_size(this->original_size_); this->finalize_data_size(); } template void Arm_input_section::do_write(Output_file* of) { // We have to write out the original section content. section_size_type section_size; const unsigned char* section_contents = this->relobj()->section_contents(this->shndx(), §ion_size, false); of->write(this->offset(), section_contents, section_size); // If this owns a stub table and it is not empty, write it. if (this->is_stub_table_owner() && !this->stub_table_->empty()) this->stub_table_->write(of); } // Finalize data size. template void Arm_input_section::set_final_data_size() { // If this owns a stub table, finalize its data size as well. if (this->is_stub_table_owner()) { uint64_t address = this->address(); // The stub table comes after the original section contents. address += this->original_size_; address = align_address(address, this->stub_table_->addralign()); off_t offset = this->offset() + (address - this->address()); this->stub_table_->set_address_and_file_offset(address, offset); address += this->stub_table_->data_size(); gold_assert(address == this->address() + this->current_data_size()); } this->set_data_size(this->current_data_size()); } // Reset address and file offset. template void Arm_input_section::do_reset_address_and_file_offset() { // Size of the original input section contents. off_t off = convert_types(this->original_size_); // If this is a stub table owner, account for the stub table size. if (this->is_stub_table_owner()) { Stub_table* stub_table = this->stub_table_; // Reset the stub table's address and file offset. The // current data size for child will be updated after that. stub_table_->reset_address_and_file_offset(); off = align_address(off, stub_table_->addralign()); off += stub_table->current_data_size(); } this->set_current_data_size(off); } // Arm_exidx_cantunwind methods. // Write this to Output file OF for a fixed endianity. template void Arm_exidx_cantunwind::do_fixed_endian_write(Output_file* of) { off_t offset = this->offset(); const section_size_type oview_size = 8; unsigned char* const oview = of->get_output_view(offset, oview_size); typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(oview); Output_section* os = this->relobj_->output_section(this->shndx_); gold_assert(os != NULL); Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(this->relobj_); Arm_address output_offset = arm_relobj->get_output_section_offset(this->shndx_); Arm_address section_start; if(output_offset != Arm_relobj::invalid_address) section_start = os->address() + output_offset; else { // Currently this only happens for a relaxed section. const Output_relaxed_input_section* poris = os->find_relaxed_input_section(this->relobj_, this->shndx_); gold_assert(poris != NULL); section_start = poris->address(); } // We always append this to the end of an EXIDX section. Arm_address output_address = section_start + this->relobj_->section_size(this->shndx_); // Write out the entry. The first word either points to the beginning // or after the end of a text section. The second word is the special // EXIDX_CANTUNWIND value. uint32_t prel31_offset = output_address - this->address(); if (utils::has_overflow<31>(offset)) gold_error(_("PREL31 overflow in EXIDX_CANTUNWIND entry")); elfcpp::Swap<32, big_endian>::writeval(wv, prel31_offset & 0x7fffffffU); elfcpp::Swap<32, big_endian>::writeval(wv + 1, elfcpp::EXIDX_CANTUNWIND); of->write_output_view(this->offset(), oview_size, oview); } // Arm_exidx_merged_section methods. // Constructor for Arm_exidx_merged_section. // EXIDX_INPUT_SECTION points to the unmodified EXIDX input section. // SECTION_OFFSET_MAP points to a section offset map describing how // parts of the input section are mapped to output. DELETED_BYTES is // the number of bytes deleted from the EXIDX input section. Arm_exidx_merged_section::Arm_exidx_merged_section( const Arm_exidx_input_section& exidx_input_section, const Arm_exidx_section_offset_map& section_offset_map, uint32_t deleted_bytes) : Output_relaxed_input_section(exidx_input_section.relobj(), exidx_input_section.shndx(), exidx_input_section.addralign()), exidx_input_section_(exidx_input_section), section_offset_map_(section_offset_map) { // Fix size here so that we do not need to implement set_final_data_size. this->set_data_size(exidx_input_section.size() - deleted_bytes); this->fix_data_size(); } // Given an input OBJECT, an input section index SHNDX within that // object, and an OFFSET relative to the start of that input // section, return whether or not the corresponding offset within // the output section is known. If this function returns true, it // sets *POUTPUT to the output offset. The value -1 indicates that // this input offset is being discarded. bool Arm_exidx_merged_section::do_output_offset( const Relobj* relobj, unsigned int shndx, section_offset_type offset, section_offset_type* poutput) const { // We only handle offsets for the original EXIDX input section. if (relobj != this->exidx_input_section_.relobj() || shndx != this->exidx_input_section_.shndx()) return false; section_offset_type section_size = convert_types(this->exidx_input_section_.size()); if (offset < 0 || offset >= section_size) // Input offset is out of valid range. *poutput = -1; else { // We need to look up the section offset map to determine the output // offset. Find the reference point in map that is first offset // bigger than or equal to this offset. Arm_exidx_section_offset_map::const_iterator p = this->section_offset_map_.lower_bound(offset); // The section offset maps are build such that this should not happen if // input offset is in the valid range. gold_assert(p != this->section_offset_map_.end()); // We need to check if this is dropped. section_offset_type ref = p->first; section_offset_type mapped_ref = p->second; if (mapped_ref != Arm_exidx_input_section::invalid_offset) // Offset is present in output. *poutput = mapped_ref + (offset - ref); else // Offset is discarded owing to EXIDX entry merging. *poutput = -1; } return true; } // Write this to output file OF. void Arm_exidx_merged_section::do_write(Output_file* of) { // If we retain or discard the whole EXIDX input section, we would // not be here. gold_assert(this->data_size() != this->exidx_input_section_.size() && this->data_size() != 0); off_t offset = this->offset(); const section_size_type oview_size = this->data_size(); unsigned char* const oview = of->get_output_view(offset, oview_size); Output_section* os = this->relobj()->output_section(this->shndx()); gold_assert(os != NULL); // Get contents of EXIDX input section. section_size_type section_size; const unsigned char* section_contents = this->relobj()->section_contents(this->shndx(), §ion_size, false); gold_assert(section_size == this->exidx_input_section_.size()); // Go over spans of input offsets and write only those that are not // discarded. section_offset_type in_start = 0; section_offset_type out_start = 0; for(Arm_exidx_section_offset_map::const_iterator p = this->section_offset_map_.begin(); p != this->section_offset_map_.end(); ++p) { section_offset_type in_end = p->first; gold_assert(in_end >= in_start); section_offset_type out_end = p->second; size_t in_chunk_size = convert_types(in_end - in_start + 1); if (out_end != -1) { size_t out_chunk_size = convert_types(out_end - out_start + 1); gold_assert(out_chunk_size == in_chunk_size); memcpy(oview + out_start, section_contents + in_start, out_chunk_size); out_start += out_chunk_size; } in_start += in_chunk_size; } gold_assert(convert_to_section_size_type(out_start) == oview_size); of->write_output_view(this->offset(), oview_size, oview); } // Arm_exidx_fixup methods. // Append an EXIDX_CANTUNWIND in the current output section if the last entry // is not an EXIDX_CANTUNWIND entry already. The new EXIDX_CANTUNWIND entry // points to the end of the last seen EXIDX section. void Arm_exidx_fixup::add_exidx_cantunwind_as_needed() { if (this->last_unwind_type_ != UT_EXIDX_CANTUNWIND && this->last_input_section_ != NULL) { Relobj* relobj = this->last_input_section_->relobj(); unsigned int text_shndx = this->last_input_section_->link(); Arm_exidx_cantunwind* cantunwind = new Arm_exidx_cantunwind(relobj, text_shndx); this->exidx_output_section_->add_output_section_data(cantunwind); this->last_unwind_type_ = UT_EXIDX_CANTUNWIND; } } // Process an EXIDX section entry in input. Return whether this entry // can be deleted in the output. SECOND_WORD in the second word of the // EXIDX entry. bool Arm_exidx_fixup::process_exidx_entry(uint32_t second_word) { bool delete_entry; if (second_word == elfcpp::EXIDX_CANTUNWIND) { // Merge if previous entry is also an EXIDX_CANTUNWIND. delete_entry = this->last_unwind_type_ == UT_EXIDX_CANTUNWIND; this->last_unwind_type_ = UT_EXIDX_CANTUNWIND; } else if ((second_word & 0x80000000) != 0) { // Inlined unwinding data. Merge if equal to previous. delete_entry = (this->last_unwind_type_ == UT_INLINED_ENTRY && this->last_inlined_entry_ == second_word); this->last_unwind_type_ = UT_INLINED_ENTRY; this->last_inlined_entry_ = second_word; } else { // Normal table entry. In theory we could merge these too, // but duplicate entries are likely to be much less common. delete_entry = false; this->last_unwind_type_ = UT_NORMAL_ENTRY; } return delete_entry; } // Update the current section offset map during EXIDX section fix-up. // If there is no map, create one. INPUT_OFFSET is the offset of a // reference point, DELETED_BYTES is the number of deleted by in the // section so far. If DELETE_ENTRY is true, the reference point and // all offsets after the previous reference point are discarded. void Arm_exidx_fixup::update_offset_map( section_offset_type input_offset, section_size_type deleted_bytes, bool delete_entry) { if (this->section_offset_map_ == NULL) this->section_offset_map_ = new Arm_exidx_section_offset_map(); section_offset_type output_offset = (delete_entry ? -1 : input_offset - deleted_bytes); (*this->section_offset_map_)[input_offset] = output_offset; } // Process EXIDX_INPUT_SECTION for EXIDX entry merging. Return the number of // bytes deleted. If some entries are merged, also store a pointer to a newly // created Arm_exidx_section_offset_map object in *PSECTION_OFFSET_MAP. The // caller owns the map and is responsible for releasing it after use. template uint32_t Arm_exidx_fixup::process_exidx_section( const Arm_exidx_input_section* exidx_input_section, Arm_exidx_section_offset_map** psection_offset_map) { Relobj* relobj = exidx_input_section->relobj(); unsigned shndx = exidx_input_section->shndx(); section_size_type section_size; const unsigned char* section_contents = relobj->section_contents(shndx, §ion_size, false); if ((section_size % 8) != 0) { // Something is wrong with this section. Better not touch it. gold_error(_("uneven .ARM.exidx section size in %s section %u"), relobj->name().c_str(), shndx); this->last_input_section_ = exidx_input_section; this->last_unwind_type_ = UT_NONE; return 0; } uint32_t deleted_bytes = 0; bool prev_delete_entry = false; gold_assert(this->section_offset_map_ == NULL); for (section_size_type i = 0; i < section_size; i += 8) { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(section_contents + i + 4); uint32_t second_word = elfcpp::Swap<32, big_endian>::readval(wv); bool delete_entry = this->process_exidx_entry(second_word); // Entry deletion causes changes in output offsets. We use a std::map // to record these. And entry (x, y) means input offset x // is mapped to output offset y. If y is invalid_offset, then x is // dropped in the output. Because of the way std::map::lower_bound // works, we record the last offset in a region w.r.t to keeping or // dropping. If there is no entry (x0, y0) for an input offset x0, // the output offset y0 of it is determined by the output offset y1 of // the smallest input offset x1 > x0 that there is an (x1, y1) entry // in the map. If y1 is not -1, then y0 = y1 + x0 - x1. Othewise, y1 // y0 is also -1. if (delete_entry != prev_delete_entry && i != 0) this->update_offset_map(i - 1, deleted_bytes, prev_delete_entry); // Update total deleted bytes for this entry. if (delete_entry) deleted_bytes += 8; prev_delete_entry = delete_entry; } // If section offset map is not NULL, make an entry for the end of // section. if (this->section_offset_map_ != NULL) update_offset_map(section_size - 1, deleted_bytes, prev_delete_entry); *psection_offset_map = this->section_offset_map_; this->section_offset_map_ = NULL; this->last_input_section_ = exidx_input_section; // Set the first output text section so that we can link the EXIDX output // section to it. Ignore any EXIDX input section that is completely merged. if (this->first_output_text_section_ == NULL && deleted_bytes != section_size) { unsigned int link = exidx_input_section->link(); Output_section* os = relobj->output_section(link); gold_assert(os != NULL); this->first_output_text_section_ = os; } return deleted_bytes; } // Arm_output_section methods. // Create a stub group for input sections from BEGIN to END. OWNER // points to the input section to be the owner a new stub table. template void Arm_output_section::create_stub_group( Input_section_list::const_iterator begin, Input_section_list::const_iterator end, Input_section_list::const_iterator owner, Target_arm* target, std::vector* new_relaxed_sections) { // We use a different kind of relaxed section in an EXIDX section. // The static casting from Output_relaxed_input_section to // Arm_input_section is invalid in an EXIDX section. We are okay // because we should not be calling this for an EXIDX section. gold_assert(this->type() != elfcpp::SHT_ARM_EXIDX); // Currently we convert ordinary input sections into relaxed sections only // at this point but we may want to support creating relaxed input section // very early. So we check here to see if owner is already a relaxed // section. Arm_input_section* arm_input_section; if (owner->is_relaxed_input_section()) { arm_input_section = Arm_input_section::as_arm_input_section( owner->relaxed_input_section()); } else { gold_assert(owner->is_input_section()); // Create a new relaxed input section. arm_input_section = target->new_arm_input_section(owner->relobj(), owner->shndx()); new_relaxed_sections->push_back(arm_input_section); } // Create a stub table. Stub_table* stub_table = target->new_stub_table(arm_input_section); arm_input_section->set_stub_table(stub_table); Input_section_list::const_iterator p = begin; Input_section_list::const_iterator prev_p; // Look for input sections or relaxed input sections in [begin ... end]. do { if (p->is_input_section() || p->is_relaxed_input_section()) { // The stub table information for input sections live // in their objects. Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(p->relobj()); arm_relobj->set_stub_table(p->shndx(), stub_table); } prev_p = p++; } while (prev_p != end); } // Group input sections for stub generation. GROUP_SIZE is roughly the limit // of stub groups. We grow a stub group by adding input section until the // size is just below GROUP_SIZE. The last input section will be converted // into a stub table. If STUB_ALWAYS_AFTER_BRANCH is false, we also add // input section after the stub table, effectively double the group size. // // This is similar to the group_sections() function in elf32-arm.c but is // implemented differently. template void Arm_output_section::group_sections( section_size_type group_size, bool stubs_always_after_branch, Target_arm* target) { // We only care about sections containing code. if ((this->flags() & elfcpp::SHF_EXECINSTR) == 0) return; // States for grouping. typedef enum { // No group is being built. NO_GROUP, // A group is being built but the stub table is not found yet. // We keep group a stub group until the size is just under GROUP_SIZE. // The last input section in the group will be used as the stub table. FINDING_STUB_SECTION, // A group is being built and we have already found a stub table. // We enter this state to grow a stub group by adding input section // after the stub table. This effectively doubles the group size. HAS_STUB_SECTION } State; // Any newly created relaxed sections are stored here. std::vector new_relaxed_sections; State state = NO_GROUP; section_size_type off = 0; section_size_type group_begin_offset = 0; section_size_type group_end_offset = 0; section_size_type stub_table_end_offset = 0; Input_section_list::const_iterator group_begin = this->input_sections().end(); Input_section_list::const_iterator stub_table = this->input_sections().end(); Input_section_list::const_iterator group_end = this->input_sections().end(); for (Input_section_list::const_iterator p = this->input_sections().begin(); p != this->input_sections().end(); ++p) { section_size_type section_begin_offset = align_address(off, p->addralign()); section_size_type section_end_offset = section_begin_offset + p->data_size(); // Check to see if we should group the previously seens sections. switch (state) { case NO_GROUP: break; case FINDING_STUB_SECTION: // Adding this section makes the group larger than GROUP_SIZE. if (section_end_offset - group_begin_offset >= group_size) { if (stubs_always_after_branch) { gold_assert(group_end != this->input_sections().end()); this->create_stub_group(group_begin, group_end, group_end, target, &new_relaxed_sections); state = NO_GROUP; } else { // But wait, there's more! Input sections up to // stub_group_size bytes after the stub table can be // handled by it too. state = HAS_STUB_SECTION; stub_table = group_end; stub_table_end_offset = group_end_offset; } } break; case HAS_STUB_SECTION: // Adding this section makes the post stub-section group larger // than GROUP_SIZE. if (section_end_offset - stub_table_end_offset >= group_size) { gold_assert(group_end != this->input_sections().end()); this->create_stub_group(group_begin, group_end, stub_table, target, &new_relaxed_sections); state = NO_GROUP; } break; default: gold_unreachable(); } // If we see an input section and currently there is no group, start // a new one. Skip any empty sections. if ((p->is_input_section() || p->is_relaxed_input_section()) && (p->relobj()->section_size(p->shndx()) != 0)) { if (state == NO_GROUP) { state = FINDING_STUB_SECTION; group_begin = p; group_begin_offset = section_begin_offset; } // Keep track of the last input section seen. group_end = p; group_end_offset = section_end_offset; } off = section_end_offset; } // Create a stub group for any ungrouped sections. if (state == FINDING_STUB_SECTION || state == HAS_STUB_SECTION) { gold_assert(group_end != this->input_sections().end()); this->create_stub_group(group_begin, group_end, (state == FINDING_STUB_SECTION ? group_end : stub_table), target, &new_relaxed_sections); } // Convert input section into relaxed input section in a batch. if (!new_relaxed_sections.empty()) this->convert_input_sections_to_relaxed_sections(new_relaxed_sections); // Update the section offsets for (size_t i = 0; i < new_relaxed_sections.size(); ++i) { Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj( new_relaxed_sections[i]->relobj()); unsigned int shndx = new_relaxed_sections[i]->shndx(); // Tell Arm_relobj that this input section is converted. arm_relobj->convert_input_section_to_relaxed_section(shndx); } } // Append non empty text sections in this to LIST in ascending // order of their position in this. template void Arm_output_section::append_text_sections_to_list( Text_section_list* list) { // We only care about text sections. if ((this->flags() & elfcpp::SHF_EXECINSTR) == 0) return; gold_assert((this->flags() & elfcpp::SHF_ALLOC) != 0); for (Input_section_list::const_iterator p = this->input_sections().begin(); p != this->input_sections().end(); ++p) { // We only care about plain or relaxed input sections. We also // ignore any merged sections. if ((p->is_input_section() || p->is_relaxed_input_section()) && p->data_size() != 0) list->push_back(Text_section_list::value_type(p->relobj(), p->shndx())); } } template void Arm_output_section::fix_exidx_coverage( const Text_section_list& sorted_text_sections, Symbol_table* symtab) { // We should only do this for the EXIDX output section. gold_assert(this->type() == elfcpp::SHT_ARM_EXIDX); // We don't want the relaxation loop to undo these changes, so we discard // the current saved states and take another one after the fix-up. this->discard_states(); // Remove all input sections. uint64_t address = this->address(); typedef std::list Simple_input_section_list; Simple_input_section_list input_sections; this->reset_address_and_file_offset(); this->get_input_sections(address, std::string(""), &input_sections); if (!this->input_sections().empty()) gold_error(_("Found non-EXIDX input sections in EXIDX output section")); // Go through all the known input sections and record them. typedef Unordered_set Section_id_set; Section_id_set known_input_sections; for (Simple_input_section_list::const_iterator p = input_sections.begin(); p != input_sections.end(); ++p) { // This should never happen. At this point, we should only see // plain EXIDX input sections. gold_assert(!p->is_relaxed_input_section()); known_input_sections.insert(Section_id(p->relobj(), p->shndx())); } Arm_exidx_fixup exidx_fixup(this); // Go over the sorted text sections. Section_id_set processed_input_sections; for (Text_section_list::const_iterator p = sorted_text_sections.begin(); p != sorted_text_sections.end(); ++p) { Relobj* relobj = p->first; unsigned int shndx = p->second; Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(relobj); const Arm_exidx_input_section* exidx_input_section = arm_relobj->exidx_input_section_by_link(shndx); // If this text section has no EXIDX section, force an EXIDX_CANTUNWIND // entry pointing to the end of the last seen EXIDX section. if (exidx_input_section == NULL) { exidx_fixup.add_exidx_cantunwind_as_needed(); continue; } Relobj* exidx_relobj = exidx_input_section->relobj(); unsigned int exidx_shndx = exidx_input_section->shndx(); Section_id sid(exidx_relobj, exidx_shndx); if (known_input_sections.find(sid) == known_input_sections.end()) { // This is odd. We have not seen this EXIDX input section before. // We cannot do fix-up. gold_error(_("EXIDX section %u of %s is not in EXIDX output section"), exidx_shndx, exidx_relobj->name().c_str()); exidx_fixup.add_exidx_cantunwind_as_needed(); continue; } // Fix up coverage and append input section to output data list. Arm_exidx_section_offset_map* section_offset_map = NULL; uint32_t deleted_bytes = exidx_fixup.process_exidx_section(exidx_input_section, §ion_offset_map); if (deleted_bytes == exidx_input_section->size()) { // The whole EXIDX section got merged. Remove it from output. gold_assert(section_offset_map == NULL); exidx_relobj->set_output_section(exidx_shndx, NULL); // All local symbols defined in this input section will be dropped. // We need to adjust output local symbol count. arm_relobj->set_output_local_symbol_count_needs_update(); } else if (deleted_bytes > 0) { // Some entries are merged. We need to convert this EXIDX input // section into a relaxed section. gold_assert(section_offset_map != NULL); Arm_exidx_merged_section* merged_section = new Arm_exidx_merged_section(*exidx_input_section, *section_offset_map, deleted_bytes); this->add_relaxed_input_section(merged_section); arm_relobj->convert_input_section_to_relaxed_section(exidx_shndx); // All local symbols defined in discarded portions of this input // section will be dropped. We need to adjust output local symbol // count. arm_relobj->set_output_local_symbol_count_needs_update(); } else { // Just add back the EXIDX input section. gold_assert(section_offset_map == NULL); Output_section::Simple_input_section sis(exidx_relobj, exidx_shndx); this->add_simple_input_section(sis, exidx_input_section->size(), exidx_input_section->addralign()); } processed_input_sections.insert(Section_id(exidx_relobj, exidx_shndx)); } // Insert an EXIDX_CANTUNWIND entry at the end of output if necessary. exidx_fixup.add_exidx_cantunwind_as_needed(); // Remove any known EXIDX input sections that are not processed. for (Simple_input_section_list::const_iterator p = input_sections.begin(); p != input_sections.end(); ++p) { if (processed_input_sections.find(Section_id(p->relobj(), p->shndx())) == processed_input_sections.end()) { // We only discard a known EXIDX section because its linked // text section has been folded by ICF. Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(p->relobj()); const Arm_exidx_input_section* exidx_input_section = arm_relobj->exidx_input_section_by_shndx(p->shndx()); gold_assert(exidx_input_section != NULL); unsigned int text_shndx = exidx_input_section->link(); gold_assert(symtab->is_section_folded(p->relobj(), text_shndx)); // Remove this from link. p->relobj()->set_output_section(p->shndx(), NULL); } } // Link exidx output section to the first seen output section and // set correct entry size. this->set_link_section(exidx_fixup.first_output_text_section()); this->set_entsize(8); // Make changes permanent. this->save_states(); this->set_section_offsets_need_adjustment(); } // Arm_relobj methods. // Determine if an input section is scannable for stub processing. SHDR is // the header of the section and SHNDX is the section index. OS is the output // section for the input section and SYMTAB is the global symbol table used to // look up ICF information. template bool Arm_relobj::section_is_scannable( const elfcpp::Shdr<32, big_endian>& shdr, unsigned int shndx, const Output_section* os, const Symbol_table *symtab) { // Skip any empty sections, unallocated sections or sections whose // type are not SHT_PROGBITS. if (shdr.get_sh_size() == 0 || (shdr.get_sh_flags() & elfcpp::SHF_ALLOC) == 0 || shdr.get_sh_type() != elfcpp::SHT_PROGBITS) return false; // Skip any discarded or ICF'ed sections. if (os == NULL || symtab->is_section_folded(this, shndx)) return false; // If this requires special offset handling, check to see if it is // a relaxed section. If this is not, then it is a merged section that // we cannot handle. if (this->is_output_section_offset_invalid(shndx)) { const Output_relaxed_input_section* poris = os->find_relaxed_input_section(this, shndx); if (poris == NULL) return false; } return true; } // Determine if we want to scan the SHNDX-th section for relocation stubs. // This is a helper for Arm_relobj::scan_sections_for_stubs() below. template bool Arm_relobj::section_needs_reloc_stub_scanning( const elfcpp::Shdr<32, big_endian>& shdr, const Relobj::Output_sections& out_sections, const Symbol_table *symtab, const unsigned char* pshdrs) { unsigned int sh_type = shdr.get_sh_type(); if (sh_type != elfcpp::SHT_REL && sh_type != elfcpp::SHT_RELA) return false; // Ignore empty section. off_t sh_size = shdr.get_sh_size(); if (sh_size == 0) return false; // Ignore reloc section with unexpected symbol table. The // error will be reported in the final link. if (this->adjust_shndx(shdr.get_sh_link()) != this->symtab_shndx()) return false; unsigned int reloc_size; if (sh_type == elfcpp::SHT_REL) reloc_size = elfcpp::Elf_sizes<32>::rel_size; else reloc_size = elfcpp::Elf_sizes<32>::rela_size; // Ignore reloc section with unexpected entsize or uneven size. // The error will be reported in the final link. if (reloc_size != shdr.get_sh_entsize() || sh_size % reloc_size != 0) return false; // Ignore reloc section with bad info. This error will be // reported in the final link. unsigned int index = this->adjust_shndx(shdr.get_sh_info()); if (index >= this->shnum()) return false; const unsigned int shdr_size = elfcpp::Elf_sizes<32>::shdr_size; const elfcpp::Shdr<32, big_endian> text_shdr(pshdrs + index * shdr_size); return this->section_is_scannable(text_shdr, index, out_sections[index], symtab); } // Return the output address of either a plain input section or a relaxed // input section. SHNDX is the section index. We define and use this // instead of calling Output_section::output_address because that is slow // for large output. template Arm_address Arm_relobj::simple_input_section_output_address( unsigned int shndx, Output_section* os) { if (this->is_output_section_offset_invalid(shndx)) { const Output_relaxed_input_section* poris = os->find_relaxed_input_section(this, shndx); // We do not handle merged sections here. gold_assert(poris != NULL); return poris->address(); } else return os->address() + this->get_output_section_offset(shndx); } // Determine if we want to scan the SHNDX-th section for non-relocation stubs. // This is a helper for Arm_relobj::scan_sections_for_stubs() below. template bool Arm_relobj::section_needs_cortex_a8_stub_scanning( const elfcpp::Shdr<32, big_endian>& shdr, unsigned int shndx, Output_section* os, const Symbol_table* symtab) { if (!this->section_is_scannable(shdr, shndx, os, symtab)) return false; // If the section does not cross any 4K-boundaries, it does not need to // be scanned. Arm_address address = this->simple_input_section_output_address(shndx, os); if ((address & ~0xfffU) == ((address + shdr.get_sh_size() - 1) & ~0xfffU)) return false; return true; } // Scan a section for Cortex-A8 workaround. template void Arm_relobj::scan_section_for_cortex_a8_erratum( const elfcpp::Shdr<32, big_endian>& shdr, unsigned int shndx, Output_section* os, Target_arm* arm_target) { Arm_address output_address = this->simple_input_section_output_address(shndx, os); // Get the section contents. section_size_type input_view_size = 0; const unsigned char* input_view = this->section_contents(shndx, &input_view_size, false); // We need to go through the mapping symbols to determine what to // scan. There are two reasons. First, we should look at THUMB code and // THUMB code only. Second, we only want to look at the 4K-page boundary // to speed up the scanning. // Look for the first mapping symbol in this section. It should be // at (shndx, 0). Mapping_symbol_position section_start(shndx, 0); typename Mapping_symbols_info::const_iterator p = this->mapping_symbols_info_.lower_bound(section_start); if (p == this->mapping_symbols_info_.end() || p->first != section_start) { gold_warning(_("Cortex-A8 erratum scanning failed because there " "is no mapping symbols for section %u of %s"), shndx, this->name().c_str()); return; } while (p != this->mapping_symbols_info_.end() && p->first.first == shndx) { typename Mapping_symbols_info::const_iterator next = this->mapping_symbols_info_.upper_bound(p->first); // Only scan part of a section with THUMB code. if (p->second == 't') { // Determine the end of this range. section_size_type span_start = convert_to_section_size_type(p->first.second); section_size_type span_end; if (next != this->mapping_symbols_info_.end() && next->first.first == shndx) span_end = convert_to_section_size_type(next->first.second); else span_end = convert_to_section_size_type(shdr.get_sh_size()); if (((span_start + output_address) & ~0xfffUL) != ((span_end + output_address - 1) & ~0xfffUL)) { arm_target->scan_span_for_cortex_a8_erratum(this, shndx, span_start, span_end, input_view, output_address); } } p = next; } } // Scan relocations for stub generation. template void Arm_relobj::scan_sections_for_stubs( Target_arm* arm_target, const Symbol_table* symtab, const Layout* layout) { unsigned int shnum = this->shnum(); const unsigned int shdr_size = elfcpp::Elf_sizes<32>::shdr_size; // Read the section headers. const unsigned char* pshdrs = this->get_view(this->elf_file()->shoff(), shnum * shdr_size, true, true); // To speed up processing, we set up hash tables for fast lookup of // input offsets to output addresses. this->initialize_input_to_output_maps(); const Relobj::Output_sections& out_sections(this->output_sections()); Relocate_info<32, big_endian> relinfo; relinfo.symtab = symtab; relinfo.layout = layout; relinfo.object = this; // Do relocation stubs scanning. const unsigned char* p = pshdrs + shdr_size; for (unsigned int i = 1; i < shnum; ++i, p += shdr_size) { const elfcpp::Shdr<32, big_endian> shdr(p); if (this->section_needs_reloc_stub_scanning(shdr, out_sections, symtab, pshdrs)) { unsigned int index = this->adjust_shndx(shdr.get_sh_info()); Arm_address output_offset = this->get_output_section_offset(index); Arm_address output_address; if(output_offset != invalid_address) output_address = out_sections[index]->address() + output_offset; else { // Currently this only happens for a relaxed section. const Output_relaxed_input_section* poris = out_sections[index]->find_relaxed_input_section(this, index); gold_assert(poris != NULL); output_address = poris->address(); } // Get the relocations. const unsigned char* prelocs = this->get_view(shdr.get_sh_offset(), shdr.get_sh_size(), true, false); // Get the section contents. This does work for the case in which // we modify the contents of an input section. We need to pass the // output view under such circumstances. section_size_type input_view_size = 0; const unsigned char* input_view = this->section_contents(index, &input_view_size, false); relinfo.reloc_shndx = i; relinfo.data_shndx = index; unsigned int sh_type = shdr.get_sh_type(); unsigned int reloc_size; if (sh_type == elfcpp::SHT_REL) reloc_size = elfcpp::Elf_sizes<32>::rel_size; else reloc_size = elfcpp::Elf_sizes<32>::rela_size; Output_section* os = out_sections[index]; arm_target->scan_section_for_stubs(&relinfo, sh_type, prelocs, shdr.get_sh_size() / reloc_size, os, output_offset == invalid_address, input_view, output_address, input_view_size); } } // Do Cortex-A8 erratum stubs scanning. This has to be done for a section // after its relocation section, if there is one, is processed for // relocation stubs. Merging this loop with the one above would have been // complicated since we would have had to make sure that relocation stub // scanning is done first. if (arm_target->fix_cortex_a8()) { const unsigned char* p = pshdrs + shdr_size; for (unsigned int i = 1; i < shnum; ++i, p += shdr_size) { const elfcpp::Shdr<32, big_endian> shdr(p); if (this->section_needs_cortex_a8_stub_scanning(shdr, i, out_sections[i], symtab)) this->scan_section_for_cortex_a8_erratum(shdr, i, out_sections[i], arm_target); } } // After we've done the relocations, we release the hash tables, // since we no longer need them. this->free_input_to_output_maps(); } // Count the local symbols. The ARM backend needs to know if a symbol // is a THUMB function or not. For global symbols, it is easy because // the Symbol object keeps the ELF symbol type. For local symbol it is // harder because we cannot access this information. So we override the // do_count_local_symbol in parent and scan local symbols to mark // THUMB functions. This is not the most efficient way but I do not want to // slow down other ports by calling a per symbol targer hook inside // Sized_relobj::do_count_local_symbols. template void Arm_relobj::do_count_local_symbols( Stringpool_template* pool, Stringpool_template* dynpool) { // We need to fix-up the values of any local symbols whose type are // STT_ARM_TFUNC. // Ask parent to count the local symbols. Sized_relobj<32, big_endian>::do_count_local_symbols(pool, dynpool); const unsigned int loccount = this->local_symbol_count(); if (loccount == 0) return; // Intialize the thumb function bit-vector. std::vector empty_vector(loccount, false); this->local_symbol_is_thumb_function_.swap(empty_vector); // Read the symbol table section header. const unsigned int symtab_shndx = this->symtab_shndx(); elfcpp::Shdr<32, big_endian> symtabshdr(this, this->elf_file()->section_header(symtab_shndx)); gold_assert(symtabshdr.get_sh_type() == elfcpp::SHT_SYMTAB); // Read the local symbols. const int sym_size =elfcpp::Elf_sizes<32>::sym_size; gold_assert(loccount == symtabshdr.get_sh_info()); off_t locsize = loccount * sym_size; const unsigned char* psyms = this->get_view(symtabshdr.get_sh_offset(), locsize, true, true); // For mapping symbol processing, we need to read the symbol names. unsigned int strtab_shndx = this->adjust_shndx(symtabshdr.get_sh_link()); if (strtab_shndx >= this->shnum()) { this->error(_("invalid symbol table name index: %u"), strtab_shndx); return; } elfcpp::Shdr<32, big_endian> strtabshdr(this, this->elf_file()->section_header(strtab_shndx)); if (strtabshdr.get_sh_type() != elfcpp::SHT_STRTAB) { this->error(_("symbol table name section has wrong type: %u"), static_cast(strtabshdr.get_sh_type())); return; } const char* pnames = reinterpret_cast(this->get_view(strtabshdr.get_sh_offset(), strtabshdr.get_sh_size(), false, false)); // Loop over the local symbols and mark any local symbols pointing // to THUMB functions. // Skip the first dummy symbol. psyms += sym_size; typename Sized_relobj<32, big_endian>::Local_values* plocal_values = this->local_values(); for (unsigned int i = 1; i < loccount; ++i, psyms += sym_size) { elfcpp::Sym<32, big_endian> sym(psyms); elfcpp::STT st_type = sym.get_st_type(); Symbol_value<32>& lv((*plocal_values)[i]); Arm_address input_value = lv.input_value(); // Check to see if this is a mapping symbol. const char* sym_name = pnames + sym.get_st_name(); if (Target_arm::is_mapping_symbol_name(sym_name)) { unsigned int input_shndx = sym.get_st_shndx(); // Strip of LSB in case this is a THUMB symbol. Mapping_symbol_position msp(input_shndx, input_value & ~1U); this->mapping_symbols_info_[msp] = sym_name[1]; } if (st_type == elfcpp::STT_ARM_TFUNC || (st_type == elfcpp::STT_FUNC && ((input_value & 1) != 0))) { // This is a THUMB function. Mark this and canonicalize the // symbol value by setting LSB. this->local_symbol_is_thumb_function_[i] = true; if ((input_value & 1) == 0) lv.set_input_value(input_value | 1); } } } // Relocate sections. template void Arm_relobj::do_relocate_sections( const Symbol_table* symtab, const Layout* layout, const unsigned char* pshdrs, typename Sized_relobj<32, big_endian>::Views* pviews) { // Call parent to relocate sections. Sized_relobj<32, big_endian>::do_relocate_sections(symtab, layout, pshdrs, pviews); // We do not generate stubs if doing a relocatable link. if (parameters->options().relocatable()) return; // Relocate stub tables. unsigned int shnum = this->shnum(); Target_arm* arm_target = Target_arm::default_target(); Relocate_info<32, big_endian> relinfo; relinfo.symtab = symtab; relinfo.layout = layout; relinfo.object = this; for (unsigned int i = 1; i < shnum; ++i) { Arm_input_section* arm_input_section = arm_target->find_arm_input_section(this, i); if (arm_input_section != NULL && arm_input_section->is_stub_table_owner() && !arm_input_section->stub_table()->empty()) { // We cannot discard a section if it owns a stub table. Output_section* os = this->output_section(i); gold_assert(os != NULL); relinfo.reloc_shndx = elfcpp::SHN_UNDEF; relinfo.reloc_shdr = NULL; relinfo.data_shndx = i; relinfo.data_shdr = pshdrs + i * elfcpp::Elf_sizes<32>::shdr_size; gold_assert((*pviews)[i].view != NULL); // We are passed the output section view. Adjust it to cover the // stub table only. Stub_table* stub_table = arm_input_section->stub_table(); gold_assert((stub_table->address() >= (*pviews)[i].address) && ((stub_table->address() + stub_table->data_size()) <= (*pviews)[i].address + (*pviews)[i].view_size)); off_t offset = stub_table->address() - (*pviews)[i].address; unsigned char* view = (*pviews)[i].view + offset; Arm_address address = stub_table->address(); section_size_type view_size = stub_table->data_size(); stub_table->relocate_stubs(&relinfo, arm_target, os, view, address, view_size); } // Apply Cortex A8 workaround if applicable. if (this->section_has_cortex_a8_workaround(i)) { unsigned char* view = (*pviews)[i].view; Arm_address view_address = (*pviews)[i].address; section_size_type view_size = (*pviews)[i].view_size; Stub_table* stub_table = this->stub_tables_[i]; // Adjust view to cover section. Output_section* os = this->output_section(i); gold_assert(os != NULL); Arm_address section_address = this->simple_input_section_output_address(i, os); uint64_t section_size = this->section_size(i); gold_assert(section_address >= view_address && ((section_address + section_size) <= (view_address + view_size))); unsigned char* section_view = view + (section_address - view_address); // Apply the Cortex-A8 workaround to the output address range // corresponding to this input section. stub_table->apply_cortex_a8_workaround_to_address_range( arm_target, section_view, section_address, section_size); } } } // Create a new EXIDX input section object for EXIDX section SHNDX with // header SHDR. template void Arm_relobj::make_exidx_input_section( unsigned int shndx, const elfcpp::Shdr<32, big_endian>& shdr) { // Link .text section to its .ARM.exidx section in the same object. unsigned int text_shndx = this->adjust_shndx(shdr.get_sh_link()); // Issue an error and ignore this EXIDX section if it does not point // to any text section. if (text_shndx == elfcpp::SHN_UNDEF) { gold_error(_("EXIDX section %u in %s has no linked text section"), shndx, this->name().c_str()); return; } // Issue an error and ignore this EXIDX section if it points to a text // section already has an EXIDX section. if (this->exidx_section_map_[text_shndx] != NULL) { gold_error(_("EXIDX sections %u and %u both link to text section %u " "in %s"), shndx, this->exidx_section_map_[text_shndx]->shndx(), text_shndx, this->name().c_str()); return; } // Create an Arm_exidx_input_section object for this EXIDX section. Arm_exidx_input_section* exidx_input_section = new Arm_exidx_input_section(this, shndx, text_shndx, shdr.get_sh_size(), shdr.get_sh_addralign()); this->exidx_section_map_[text_shndx] = exidx_input_section; // Also map the EXIDX section index to this. gold_assert(this->exidx_section_map_[shndx] == NULL); this->exidx_section_map_[shndx] = exidx_input_section; } // Read the symbol information. template void Arm_relobj::do_read_symbols(Read_symbols_data* sd) { // Call parent class to read symbol information. Sized_relobj<32, big_endian>::do_read_symbols(sd); // Read processor-specific flags in ELF file header. const unsigned char* pehdr = this->get_view(elfcpp::file_header_offset, elfcpp::Elf_sizes<32>::ehdr_size, true, false); elfcpp::Ehdr<32, big_endian> ehdr(pehdr); this->processor_specific_flags_ = ehdr.get_e_flags(); // Go over the section headers and look for .ARM.attributes and .ARM.exidx // sections. const size_t shdr_size = elfcpp::Elf_sizes<32>::shdr_size; const unsigned char *ps = sd->section_headers->data() + shdr_size; for (unsigned int i = 1; i < this->shnum(); ++i, ps += shdr_size) { elfcpp::Shdr<32, big_endian> shdr(ps); if (shdr.get_sh_type() == elfcpp::SHT_ARM_ATTRIBUTES) { gold_assert(this->attributes_section_data_ == NULL); section_offset_type section_offset = shdr.get_sh_offset(); section_size_type section_size = convert_to_section_size_type(shdr.get_sh_size()); File_view* view = this->get_lasting_view(section_offset, section_size, true, false); this->attributes_section_data_ = new Attributes_section_data(view->data(), section_size); } else if (shdr.get_sh_type() == elfcpp::SHT_ARM_EXIDX) this->make_exidx_input_section(i, shdr); } } // Process relocations for garbage collection. The ARM target uses .ARM.exidx // sections for unwinding. These sections are referenced implicitly by // text sections linked in the section headers. If we ignore these implict // references, the .ARM.exidx sections and any .ARM.extab sections they use // will be garbage-collected incorrectly. Hence we override the same function // in the base class to handle these implicit references. template void Arm_relobj::do_gc_process_relocs(Symbol_table* symtab, Layout* layout, Read_relocs_data* rd) { // First, call base class method to process relocations in this object. Sized_relobj<32, big_endian>::do_gc_process_relocs(symtab, layout, rd); unsigned int shnum = this->shnum(); const unsigned int shdr_size = elfcpp::Elf_sizes<32>::shdr_size; const unsigned char* pshdrs = this->get_view(this->elf_file()->shoff(), shnum * shdr_size, true, true); // Scan section headers for sections of type SHT_ARM_EXIDX. Add references // to these from the linked text sections. const unsigned char* ps = pshdrs + shdr_size; for (unsigned int i = 1; i < shnum; ++i, ps += shdr_size) { elfcpp::Shdr<32, big_endian> shdr(ps); if (shdr.get_sh_type() == elfcpp::SHT_ARM_EXIDX) { // Found an .ARM.exidx section, add it to the set of reachable // sections from its linked text section. unsigned int text_shndx = this->adjust_shndx(shdr.get_sh_link()); symtab->gc()->add_reference(this, text_shndx, this, i); } } } // Update output local symbol count. Owing to EXIDX entry merging, some local // symbols will be removed in output. Adjust output local symbol count // accordingly. We can only changed the static output local symbol count. It // is too late to change the dynamic symbols. template void Arm_relobj::update_output_local_symbol_count() { // Caller should check that this needs updating. We want caller checking // because output_local_symbol_count_needs_update() is most likely inlined. gold_assert(this->output_local_symbol_count_needs_update_); gold_assert(this->symtab_shndx() != -1U); if (this->symtab_shndx() == 0) { // This object has no symbols. Weird but legal. return; } // Read the symbol table section header. const unsigned int symtab_shndx = this->symtab_shndx(); elfcpp::Shdr<32, big_endian> symtabshdr(this, this->elf_file()->section_header(symtab_shndx)); gold_assert(symtabshdr.get_sh_type() == elfcpp::SHT_SYMTAB); // Read the local symbols. const int sym_size = elfcpp::Elf_sizes<32>::sym_size; const unsigned int loccount = this->local_symbol_count(); gold_assert(loccount == symtabshdr.get_sh_info()); off_t locsize = loccount * sym_size; const unsigned char* psyms = this->get_view(symtabshdr.get_sh_offset(), locsize, true, true); // Loop over the local symbols. typedef typename Sized_relobj<32, big_endian>::Output_sections Output_sections; const Output_sections& out_sections(this->output_sections()); unsigned int shnum = this->shnum(); unsigned int count = 0; // Skip the first, dummy, symbol. psyms += sym_size; for (unsigned int i = 1; i < loccount; ++i, psyms += sym_size) { elfcpp::Sym<32, big_endian> sym(psyms); Symbol_value<32>& lv((*this->local_values())[i]); // This local symbol was already discarded by do_count_local_symbols. if (!lv.needs_output_symtab_entry()) continue; bool is_ordinary; unsigned int shndx = this->adjust_sym_shndx(i, sym.get_st_shndx(), &is_ordinary); if (shndx < shnum) { Output_section* os = out_sections[shndx]; // This local symbol no longer has an output section. Discard it. if (os == NULL) { lv.set_no_output_symtab_entry(); continue; } // Currently we only discard parts of EXIDX input sections. // We explicitly check for a merged EXIDX input section to avoid // calling Output_section_data::output_offset unless necessary. if ((this->get_output_section_offset(shndx) == invalid_address) && (this->exidx_input_section_by_shndx(shndx) != NULL)) { section_offset_type output_offset = os->output_offset(this, shndx, lv.input_value()); if (output_offset == -1) { // This symbol is defined in a part of an EXIDX input section // that is discarded due to entry merging. lv.set_no_output_symtab_entry(); continue; } } } ++count; } this->set_output_local_symbol_count(count); this->output_local_symbol_count_needs_update_ = false; } // Arm_dynobj methods. // Read the symbol information. template void Arm_dynobj::do_read_symbols(Read_symbols_data* sd) { // Call parent class to read symbol information. Sized_dynobj<32, big_endian>::do_read_symbols(sd); // Read processor-specific flags in ELF file header. const unsigned char* pehdr = this->get_view(elfcpp::file_header_offset, elfcpp::Elf_sizes<32>::ehdr_size, true, false); elfcpp::Ehdr<32, big_endian> ehdr(pehdr); this->processor_specific_flags_ = ehdr.get_e_flags(); // Read the attributes section if there is one. // We read from the end because gas seems to put it near the end of // the section headers. const size_t shdr_size = elfcpp::Elf_sizes<32>::shdr_size; const unsigned char *ps = sd->section_headers->data() + shdr_size * (this->shnum() - 1); for (unsigned int i = this->shnum(); i > 0; --i, ps -= shdr_size) { elfcpp::Shdr<32, big_endian> shdr(ps); if (shdr.get_sh_type() == elfcpp::SHT_ARM_ATTRIBUTES) { section_offset_type section_offset = shdr.get_sh_offset(); section_size_type section_size = convert_to_section_size_type(shdr.get_sh_size()); File_view* view = this->get_lasting_view(section_offset, section_size, true, false); this->attributes_section_data_ = new Attributes_section_data(view->data(), section_size); break; } } } // Stub_addend_reader methods. // Read the addend of a REL relocation of type R_TYPE at VIEW. template elfcpp::Elf_types<32>::Elf_Swxword Stub_addend_reader::operator()( unsigned int r_type, const unsigned char* view, const typename Reloc_types::Reloc&) const { typedef struct Arm_relocate_functions RelocFuncs; switch (r_type) { case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_PLT32: { typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(view); Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); return utils::sign_extend<26>(val << 2); } case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_THM_JUMP24: case elfcpp::R_ARM_THM_XPC22: { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(view); Valtype upper_insn = elfcpp::Swap<16, big_endian>::readval(wv); Valtype lower_insn = elfcpp::Swap<16, big_endian>::readval(wv + 1); return RelocFuncs::thumb32_branch_offset(upper_insn, lower_insn); } case elfcpp::R_ARM_THM_JUMP19: { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(view); Valtype upper_insn = elfcpp::Swap<16, big_endian>::readval(wv); Valtype lower_insn = elfcpp::Swap<16, big_endian>::readval(wv + 1); return RelocFuncs::thumb32_cond_branch_offset(upper_insn, lower_insn); } default: gold_unreachable(); } } // A class to handle the PLT data. template class Output_data_plt_arm : public Output_section_data { public: typedef Output_data_reloc Reloc_section; Output_data_plt_arm(Layout*, Output_data_space*); // Add an entry to the PLT. void add_entry(Symbol* gsym); // Return the .rel.plt section data. const Reloc_section* rel_plt() const { return this->rel_; } protected: void do_adjust_output_section(Output_section* os); // Write to a map file. void do_print_to_mapfile(Mapfile* mapfile) const { mapfile->print_output_data(this, _("** PLT")); } private: // Template for the first PLT entry. static const uint32_t first_plt_entry[5]; // Template for subsequent PLT entries. static const uint32_t plt_entry[3]; // Set the final size. void set_final_data_size() { this->set_data_size(sizeof(first_plt_entry) + this->count_ * sizeof(plt_entry)); } // Write out the PLT data. void do_write(Output_file*); // The reloc section. Reloc_section* rel_; // The .got.plt section. Output_data_space* got_plt_; // The number of PLT entries. unsigned int count_; }; // Create the PLT section. The ordinary .got section is an argument, // since we need to refer to the start. We also create our own .got // section just for PLT entries. template Output_data_plt_arm::Output_data_plt_arm(Layout* layout, Output_data_space* got_plt) : Output_section_data(4), got_plt_(got_plt), count_(0) { this->rel_ = new Reloc_section(false); layout->add_output_section_data(".rel.plt", elfcpp::SHT_REL, elfcpp::SHF_ALLOC, this->rel_, true, false, false, false); } template void Output_data_plt_arm::do_adjust_output_section(Output_section* os) { os->set_entsize(0); } // Add an entry to the PLT. template void Output_data_plt_arm::add_entry(Symbol* gsym) { gold_assert(!gsym->has_plt_offset()); // Note that when setting the PLT offset we skip the initial // reserved PLT entry. gsym->set_plt_offset((this->count_) * sizeof(plt_entry) + sizeof(first_plt_entry)); ++this->count_; section_offset_type got_offset = this->got_plt_->current_data_size(); // Every PLT entry needs a GOT entry which points back to the PLT // entry (this will be changed by the dynamic linker, normally // lazily when the function is called). this->got_plt_->set_current_data_size(got_offset + 4); // Every PLT entry needs a reloc. gsym->set_needs_dynsym_entry(); this->rel_->add_global(gsym, elfcpp::R_ARM_JUMP_SLOT, this->got_plt_, got_offset); // Note that we don't need to save the symbol. The contents of the // PLT are independent of which symbols are used. The symbols only // appear in the relocations. } // ARM PLTs. // FIXME: This is not very flexible. Right now this has only been tested // on armv5te. If we are to support additional architecture features like // Thumb-2 or BE8, we need to make this more flexible like GNU ld. // The first entry in the PLT. template const uint32_t Output_data_plt_arm::first_plt_entry[5] = { 0xe52de004, // str lr, [sp, #-4]! 0xe59fe004, // ldr lr, [pc, #4] 0xe08fe00e, // add lr, pc, lr 0xe5bef008, // ldr pc, [lr, #8]! 0x00000000, // &GOT[0] - . }; // Subsequent entries in the PLT. template const uint32_t Output_data_plt_arm::plt_entry[3] = { 0xe28fc600, // add ip, pc, #0xNN00000 0xe28cca00, // add ip, ip, #0xNN000 0xe5bcf000, // ldr pc, [ip, #0xNNN]! }; // Write out the PLT. This uses the hand-coded instructions above, // and adjusts them as needed. This is all specified by the arm ELF // Processor Supplement. template void Output_data_plt_arm::do_write(Output_file* of) { const off_t offset = this->offset(); const section_size_type oview_size = convert_to_section_size_type(this->data_size()); unsigned char* const oview = of->get_output_view(offset, oview_size); const off_t got_file_offset = this->got_plt_->offset(); const section_size_type got_size = convert_to_section_size_type(this->got_plt_->data_size()); unsigned char* const got_view = of->get_output_view(got_file_offset, got_size); unsigned char* pov = oview; Arm_address plt_address = this->address(); Arm_address got_address = this->got_plt_->address(); // Write first PLT entry. All but the last word are constants. const size_t num_first_plt_words = (sizeof(first_plt_entry) / sizeof(plt_entry[0])); for (size_t i = 0; i < num_first_plt_words - 1; i++) elfcpp::Swap<32, big_endian>::writeval(pov + i * 4, first_plt_entry[i]); // Last word in first PLT entry is &GOT[0] - . elfcpp::Swap<32, big_endian>::writeval(pov + 16, got_address - (plt_address + 16)); pov += sizeof(first_plt_entry); unsigned char* got_pov = got_view; memset(got_pov, 0, 12); got_pov += 12; const int rel_size = elfcpp::Elf_sizes<32>::rel_size; unsigned int plt_offset = sizeof(first_plt_entry); unsigned int plt_rel_offset = 0; unsigned int got_offset = 12; const unsigned int count = this->count_; for (unsigned int i = 0; i < count; ++i, pov += sizeof(plt_entry), got_pov += 4, plt_offset += sizeof(plt_entry), plt_rel_offset += rel_size, got_offset += 4) { // Set and adjust the PLT entry itself. int32_t offset = ((got_address + got_offset) - (plt_address + plt_offset + 8)); gold_assert(offset >= 0 && offset < 0x0fffffff); uint32_t plt_insn0 = plt_entry[0] | ((offset >> 20) & 0xff); elfcpp::Swap<32, big_endian>::writeval(pov, plt_insn0); uint32_t plt_insn1 = plt_entry[1] | ((offset >> 12) & 0xff); elfcpp::Swap<32, big_endian>::writeval(pov + 4, plt_insn1); uint32_t plt_insn2 = plt_entry[2] | (offset & 0xfff); elfcpp::Swap<32, big_endian>::writeval(pov + 8, plt_insn2); // Set the entry in the GOT. elfcpp::Swap<32, big_endian>::writeval(got_pov, plt_address); } gold_assert(static_cast(pov - oview) == oview_size); gold_assert(static_cast(got_pov - got_view) == got_size); of->write_output_view(offset, oview_size, oview); of->write_output_view(got_file_offset, got_size, got_view); } // Create a PLT entry for a global symbol. template void Target_arm::make_plt_entry(Symbol_table* symtab, Layout* layout, Symbol* gsym) { if (gsym->has_plt_offset()) return; if (this->plt_ == NULL) { // Create the GOT sections first. this->got_section(symtab, layout); this->plt_ = new Output_data_plt_arm(layout, this->got_plt_); layout->add_output_section_data(".plt", elfcpp::SHT_PROGBITS, (elfcpp::SHF_ALLOC | elfcpp::SHF_EXECINSTR), this->plt_, false, false, false, false); } this->plt_->add_entry(gsym); } // Report an unsupported relocation against a local symbol. template void Target_arm::Scan::unsupported_reloc_local( Sized_relobj<32, big_endian>* object, unsigned int r_type) { gold_error(_("%s: unsupported reloc %u against local symbol"), object->name().c_str(), r_type); } // We are about to emit a dynamic relocation of type R_TYPE. If the // dynamic linker does not support it, issue an error. The GNU linker // only issues a non-PIC error for an allocated read-only section. // Here we know the section is allocated, but we don't know that it is // read-only. But we check for all the relocation types which the // glibc dynamic linker supports, so it seems appropriate to issue an // error even if the section is not read-only. template void Target_arm::Scan::check_non_pic(Relobj* object, unsigned int r_type) { switch (r_type) { // These are the relocation types supported by glibc for ARM. case elfcpp::R_ARM_RELATIVE: case elfcpp::R_ARM_COPY: case elfcpp::R_ARM_GLOB_DAT: case elfcpp::R_ARM_JUMP_SLOT: case elfcpp::R_ARM_ABS32: case elfcpp::R_ARM_ABS32_NOI: case elfcpp::R_ARM_PC24: // FIXME: The following 3 types are not supported by Android's dynamic // linker. case elfcpp::R_ARM_TLS_DTPMOD32: case elfcpp::R_ARM_TLS_DTPOFF32: case elfcpp::R_ARM_TLS_TPOFF32: return; default: // This prevents us from issuing more than one error per reloc // section. But we can still wind up issuing more than one // error per object file. if (this->issued_non_pic_error_) return; object->error(_("requires unsupported dynamic reloc; " "recompile with -fPIC")); this->issued_non_pic_error_ = true; return; case elfcpp::R_ARM_NONE: gold_unreachable(); } } // Scan a relocation for a local symbol. // FIXME: This only handles a subset of relocation types used by Android // on ARM v5te devices. template inline void Target_arm::Scan::local(Symbol_table* symtab, Layout* layout, Target_arm* target, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, Output_section* output_section, const elfcpp::Rel<32, big_endian>& reloc, unsigned int r_type, const elfcpp::Sym<32, big_endian>& lsym) { r_type = get_real_reloc_type(r_type); switch (r_type) { case elfcpp::R_ARM_NONE: case elfcpp::R_ARM_V4BX: case elfcpp::R_ARM_GNU_VTENTRY: case elfcpp::R_ARM_GNU_VTINHERIT: break; case elfcpp::R_ARM_ABS32: case elfcpp::R_ARM_ABS32_NOI: // If building a shared library (or a position-independent // executable), we need to create a dynamic relocation for // this location. The relocation applied at link time will // apply the link-time value, so we flag the location with // an R_ARM_RELATIVE relocation so the dynamic loader can // relocate it easily. if (parameters->options().output_is_position_independent()) { Reloc_section* rel_dyn = target->rel_dyn_section(layout); unsigned int r_sym = elfcpp::elf_r_sym<32>(reloc.get_r_info()); // If we are to add more other reloc types than R_ARM_ABS32, // we need to add check_non_pic(object, r_type) here. rel_dyn->add_local_relative(object, r_sym, elfcpp::R_ARM_RELATIVE, output_section, data_shndx, reloc.get_r_offset()); } break; case elfcpp::R_ARM_ABS16: case elfcpp::R_ARM_ABS12: case elfcpp::R_ARM_THM_ABS5: case elfcpp::R_ARM_ABS8: case elfcpp::R_ARM_BASE_ABS: case elfcpp::R_ARM_MOVW_ABS_NC: case elfcpp::R_ARM_MOVT_ABS: case elfcpp::R_ARM_THM_MOVW_ABS_NC: case elfcpp::R_ARM_THM_MOVT_ABS: // If building a shared library (or a position-independent // executable), we need to create a dynamic relocation for // this location. Because the addend needs to remain in the // data section, we need to be careful not to apply this // relocation statically. if (parameters->options().output_is_position_independent()) { check_non_pic(object, r_type); Reloc_section* rel_dyn = target->rel_dyn_section(layout); unsigned int r_sym = elfcpp::elf_r_sym<32>(reloc.get_r_info()); if (lsym.get_st_type() != elfcpp::STT_SECTION) rel_dyn->add_local(object, r_sym, r_type, output_section, data_shndx, reloc.get_r_offset()); else { gold_assert(lsym.get_st_value() == 0); unsigned int shndx = lsym.get_st_shndx(); bool is_ordinary; shndx = object->adjust_sym_shndx(r_sym, shndx, &is_ordinary); if (!is_ordinary) object->error(_("section symbol %u has bad shndx %u"), r_sym, shndx); else rel_dyn->add_local_section(object, shndx, r_type, output_section, data_shndx, reloc.get_r_offset()); } } break; case elfcpp::R_ARM_PC24: case elfcpp::R_ARM_REL32: case elfcpp::R_ARM_LDR_PC_G0: case elfcpp::R_ARM_SBREL32: case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_THM_PC8: case elfcpp::R_ARM_BASE_PREL: case elfcpp::R_ARM_PLT32: case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_THM_JUMP24: case elfcpp::R_ARM_LDR_SBREL_11_0_NC: case elfcpp::R_ARM_ALU_SBREL_19_12_NC: case elfcpp::R_ARM_ALU_SBREL_27_20_CK: case elfcpp::R_ARM_SBREL31: case elfcpp::R_ARM_PREL31: case elfcpp::R_ARM_MOVW_PREL_NC: case elfcpp::R_ARM_MOVT_PREL: case elfcpp::R_ARM_THM_MOVW_PREL_NC: case elfcpp::R_ARM_THM_MOVT_PREL: case elfcpp::R_ARM_THM_JUMP19: case elfcpp::R_ARM_THM_JUMP6: case elfcpp::R_ARM_THM_ALU_PREL_11_0: case elfcpp::R_ARM_THM_PC12: case elfcpp::R_ARM_REL32_NOI: case elfcpp::R_ARM_ALU_PC_G0_NC: case elfcpp::R_ARM_ALU_PC_G0: case elfcpp::R_ARM_ALU_PC_G1_NC: case elfcpp::R_ARM_ALU_PC_G1: case elfcpp::R_ARM_ALU_PC_G2: case elfcpp::R_ARM_LDR_PC_G1: case elfcpp::R_ARM_LDR_PC_G2: case elfcpp::R_ARM_LDRS_PC_G0: case elfcpp::R_ARM_LDRS_PC_G1: case elfcpp::R_ARM_LDRS_PC_G2: case elfcpp::R_ARM_LDC_PC_G0: case elfcpp::R_ARM_LDC_PC_G1: case elfcpp::R_ARM_LDC_PC_G2: case elfcpp::R_ARM_ALU_SB_G0_NC: case elfcpp::R_ARM_ALU_SB_G0: case elfcpp::R_ARM_ALU_SB_G1_NC: case elfcpp::R_ARM_ALU_SB_G1: case elfcpp::R_ARM_ALU_SB_G2: case elfcpp::R_ARM_LDR_SB_G0: case elfcpp::R_ARM_LDR_SB_G1: case elfcpp::R_ARM_LDR_SB_G2: case elfcpp::R_ARM_LDRS_SB_G0: case elfcpp::R_ARM_LDRS_SB_G1: case elfcpp::R_ARM_LDRS_SB_G2: case elfcpp::R_ARM_LDC_SB_G0: case elfcpp::R_ARM_LDC_SB_G1: case elfcpp::R_ARM_LDC_SB_G2: case elfcpp::R_ARM_MOVW_BREL_NC: case elfcpp::R_ARM_MOVT_BREL: case elfcpp::R_ARM_MOVW_BREL: case elfcpp::R_ARM_THM_MOVW_BREL_NC: case elfcpp::R_ARM_THM_MOVT_BREL: case elfcpp::R_ARM_THM_MOVW_BREL: case elfcpp::R_ARM_THM_JUMP11: case elfcpp::R_ARM_THM_JUMP8: // We don't need to do anything for a relative addressing relocation // against a local symbol if it does not reference the GOT. break; case elfcpp::R_ARM_GOTOFF32: case elfcpp::R_ARM_GOTOFF12: // We need a GOT section: target->got_section(symtab, layout); break; case elfcpp::R_ARM_GOT_BREL: case elfcpp::R_ARM_GOT_PREL: { // The symbol requires a GOT entry. Output_data_got<32, big_endian>* got = target->got_section(symtab, layout); unsigned int r_sym = elfcpp::elf_r_sym<32>(reloc.get_r_info()); if (got->add_local(object, r_sym, GOT_TYPE_STANDARD)) { // If we are generating a shared object, we need to add a // dynamic RELATIVE relocation for this symbol's GOT entry. if (parameters->options().output_is_position_independent()) { Reloc_section* rel_dyn = target->rel_dyn_section(layout); unsigned int r_sym = elfcpp::elf_r_sym<32>(reloc.get_r_info()); rel_dyn->add_local_relative( object, r_sym, elfcpp::R_ARM_RELATIVE, got, object->local_got_offset(r_sym, GOT_TYPE_STANDARD)); } } } break; case elfcpp::R_ARM_TARGET1: case elfcpp::R_ARM_TARGET2: // This should have been mapped to another type already. // Fall through. case elfcpp::R_ARM_COPY: case elfcpp::R_ARM_GLOB_DAT: case elfcpp::R_ARM_JUMP_SLOT: case elfcpp::R_ARM_RELATIVE: // These are relocations which should only be seen by the // dynamic linker, and should never be seen here. gold_error(_("%s: unexpected reloc %u in object file"), object->name().c_str(), r_type); break; default: unsupported_reloc_local(object, r_type); break; } } // Report an unsupported relocation against a global symbol. template void Target_arm::Scan::unsupported_reloc_global( Sized_relobj<32, big_endian>* object, unsigned int r_type, Symbol* gsym) { gold_error(_("%s: unsupported reloc %u against global symbol %s"), object->name().c_str(), r_type, gsym->demangled_name().c_str()); } // Scan a relocation for a global symbol. template inline void Target_arm::Scan::global(Symbol_table* symtab, Layout* layout, Target_arm* target, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, Output_section* output_section, const elfcpp::Rel<32, big_endian>& reloc, unsigned int r_type, Symbol* gsym) { r_type = get_real_reloc_type(r_type); switch (r_type) { case elfcpp::R_ARM_NONE: case elfcpp::R_ARM_V4BX: case elfcpp::R_ARM_GNU_VTENTRY: case elfcpp::R_ARM_GNU_VTINHERIT: break; case elfcpp::R_ARM_ABS32: case elfcpp::R_ARM_ABS16: case elfcpp::R_ARM_ABS12: case elfcpp::R_ARM_THM_ABS5: case elfcpp::R_ARM_ABS8: case elfcpp::R_ARM_BASE_ABS: case elfcpp::R_ARM_MOVW_ABS_NC: case elfcpp::R_ARM_MOVT_ABS: case elfcpp::R_ARM_THM_MOVW_ABS_NC: case elfcpp::R_ARM_THM_MOVT_ABS: case elfcpp::R_ARM_ABS32_NOI: // Absolute addressing relocations. { // Make a PLT entry if necessary. if (this->symbol_needs_plt_entry(gsym)) { target->make_plt_entry(symtab, layout, gsym); // Since this is not a PC-relative relocation, we may be // taking the address of a function. In that case we need to // set the entry in the dynamic symbol table to the address of // the PLT entry. if (gsym->is_from_dynobj() && !parameters->options().shared()) gsym->set_needs_dynsym_value(); } // Make a dynamic relocation if necessary. if (gsym->needs_dynamic_reloc(Symbol::ABSOLUTE_REF)) { if (gsym->may_need_copy_reloc()) { target->copy_reloc(symtab, layout, object, data_shndx, output_section, gsym, reloc); } else if ((r_type == elfcpp::R_ARM_ABS32 || r_type == elfcpp::R_ARM_ABS32_NOI) && gsym->can_use_relative_reloc(false)) { Reloc_section* rel_dyn = target->rel_dyn_section(layout); rel_dyn->add_global_relative(gsym, elfcpp::R_ARM_RELATIVE, output_section, object, data_shndx, reloc.get_r_offset()); } else { check_non_pic(object, r_type); Reloc_section* rel_dyn = target->rel_dyn_section(layout); rel_dyn->add_global(gsym, r_type, output_section, object, data_shndx, reloc.get_r_offset()); } } } break; case elfcpp::R_ARM_GOTOFF32: case elfcpp::R_ARM_GOTOFF12: // We need a GOT section. target->got_section(symtab, layout); break; case elfcpp::R_ARM_REL32: case elfcpp::R_ARM_LDR_PC_G0: case elfcpp::R_ARM_SBREL32: case elfcpp::R_ARM_THM_PC8: case elfcpp::R_ARM_BASE_PREL: case elfcpp::R_ARM_LDR_SBREL_11_0_NC: case elfcpp::R_ARM_ALU_SBREL_19_12_NC: case elfcpp::R_ARM_ALU_SBREL_27_20_CK: case elfcpp::R_ARM_MOVW_PREL_NC: case elfcpp::R_ARM_MOVT_PREL: case elfcpp::R_ARM_THM_MOVW_PREL_NC: case elfcpp::R_ARM_THM_MOVT_PREL: case elfcpp::R_ARM_THM_ALU_PREL_11_0: case elfcpp::R_ARM_THM_PC12: case elfcpp::R_ARM_REL32_NOI: case elfcpp::R_ARM_ALU_PC_G0_NC: case elfcpp::R_ARM_ALU_PC_G0: case elfcpp::R_ARM_ALU_PC_G1_NC: case elfcpp::R_ARM_ALU_PC_G1: case elfcpp::R_ARM_ALU_PC_G2: case elfcpp::R_ARM_LDR_PC_G1: case elfcpp::R_ARM_LDR_PC_G2: case elfcpp::R_ARM_LDRS_PC_G0: case elfcpp::R_ARM_LDRS_PC_G1: case elfcpp::R_ARM_LDRS_PC_G2: case elfcpp::R_ARM_LDC_PC_G0: case elfcpp::R_ARM_LDC_PC_G1: case elfcpp::R_ARM_LDC_PC_G2: case elfcpp::R_ARM_ALU_SB_G0_NC: case elfcpp::R_ARM_ALU_SB_G0: case elfcpp::R_ARM_ALU_SB_G1_NC: case elfcpp::R_ARM_ALU_SB_G1: case elfcpp::R_ARM_ALU_SB_G2: case elfcpp::R_ARM_LDR_SB_G0: case elfcpp::R_ARM_LDR_SB_G1: case elfcpp::R_ARM_LDR_SB_G2: case elfcpp::R_ARM_LDRS_SB_G0: case elfcpp::R_ARM_LDRS_SB_G1: case elfcpp::R_ARM_LDRS_SB_G2: case elfcpp::R_ARM_LDC_SB_G0: case elfcpp::R_ARM_LDC_SB_G1: case elfcpp::R_ARM_LDC_SB_G2: case elfcpp::R_ARM_MOVW_BREL_NC: case elfcpp::R_ARM_MOVT_BREL: case elfcpp::R_ARM_MOVW_BREL: case elfcpp::R_ARM_THM_MOVW_BREL_NC: case elfcpp::R_ARM_THM_MOVT_BREL: case elfcpp::R_ARM_THM_MOVW_BREL: // Relative addressing relocations. { // Make a dynamic relocation if necessary. int flags = Symbol::NON_PIC_REF; if (gsym->needs_dynamic_reloc(flags)) { if (target->may_need_copy_reloc(gsym)) { target->copy_reloc(symtab, layout, object, data_shndx, output_section, gsym, reloc); } else { check_non_pic(object, r_type); Reloc_section* rel_dyn = target->rel_dyn_section(layout); rel_dyn->add_global(gsym, r_type, output_section, object, data_shndx, reloc.get_r_offset()); } } } break; case elfcpp::R_ARM_PC24: case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_PLT32: case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_THM_JUMP24: case elfcpp::R_ARM_SBREL31: case elfcpp::R_ARM_PREL31: case elfcpp::R_ARM_THM_JUMP19: case elfcpp::R_ARM_THM_JUMP6: case elfcpp::R_ARM_THM_JUMP11: case elfcpp::R_ARM_THM_JUMP8: // All the relocation above are branches except for the PREL31 ones. // A PREL31 relocation can point to a personality function in a shared // library. In that case we want to use a PLT because we want to // call the personality routine and the dyanmic linkers we care about // do not support dynamic PREL31 relocations. An REL31 relocation may // point to a function whose unwinding behaviour is being described but // we will not mistakenly generate a PLT for that because we should use // a local section symbol. // If the symbol is fully resolved, this is just a relative // local reloc. Otherwise we need a PLT entry. if (gsym->final_value_is_known()) break; // If building a shared library, we can also skip the PLT entry // if the symbol is defined in the output file and is protected // or hidden. if (gsym->is_defined() && !gsym->is_from_dynobj() && !gsym->is_preemptible()) break; target->make_plt_entry(symtab, layout, gsym); break; case elfcpp::R_ARM_GOT_BREL: case elfcpp::R_ARM_GOT_ABS: case elfcpp::R_ARM_GOT_PREL: { // The symbol requires a GOT entry. Output_data_got<32, big_endian>* got = target->got_section(symtab, layout); if (gsym->final_value_is_known()) got->add_global(gsym, GOT_TYPE_STANDARD); else { // If this symbol is not fully resolved, we need to add a // GOT entry with a dynamic relocation. Reloc_section* rel_dyn = target->rel_dyn_section(layout); if (gsym->is_from_dynobj() || gsym->is_undefined() || gsym->is_preemptible()) got->add_global_with_rel(gsym, GOT_TYPE_STANDARD, rel_dyn, elfcpp::R_ARM_GLOB_DAT); else { if (got->add_global(gsym, GOT_TYPE_STANDARD)) rel_dyn->add_global_relative( gsym, elfcpp::R_ARM_RELATIVE, got, gsym->got_offset(GOT_TYPE_STANDARD)); } } } break; case elfcpp::R_ARM_TARGET1: case elfcpp::R_ARM_TARGET2: // These should have been mapped to other types already. // Fall through. case elfcpp::R_ARM_COPY: case elfcpp::R_ARM_GLOB_DAT: case elfcpp::R_ARM_JUMP_SLOT: case elfcpp::R_ARM_RELATIVE: // These are relocations which should only be seen by the // dynamic linker, and should never be seen here. gold_error(_("%s: unexpected reloc %u in object file"), object->name().c_str(), r_type); break; default: unsupported_reloc_global(object, r_type, gsym); break; } } // Process relocations for gc. template void Target_arm::gc_process_relocs(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols) { typedef Target_arm Arm; typedef typename Target_arm::Scan Scan; gold::gc_process_relocs<32, big_endian, Arm, elfcpp::SHT_REL, Scan>( symtab, layout, this, object, data_shndx, prelocs, reloc_count, output_section, needs_special_offset_handling, local_symbol_count, plocal_symbols); } // Scan relocations for a section. template void Target_arm::scan_relocs(Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols) { typedef typename Target_arm::Scan Scan; if (sh_type == elfcpp::SHT_RELA) { gold_error(_("%s: unsupported RELA reloc section"), object->name().c_str()); return; } gold::scan_relocs<32, big_endian, Target_arm, elfcpp::SHT_REL, Scan>( symtab, layout, this, object, data_shndx, prelocs, reloc_count, output_section, needs_special_offset_handling, local_symbol_count, plocal_symbols); } // Finalize the sections. template void Target_arm::do_finalize_sections( Layout* layout, const Input_objects* input_objects, Symbol_table* symtab) { // Merge processor-specific flags. for (Input_objects::Relobj_iterator p = input_objects->relobj_begin(); p != input_objects->relobj_end(); ++p) { Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(*p); this->merge_processor_specific_flags( arm_relobj->name(), arm_relobj->processor_specific_flags()); this->merge_object_attributes(arm_relobj->name().c_str(), arm_relobj->attributes_section_data()); } for (Input_objects::Dynobj_iterator p = input_objects->dynobj_begin(); p != input_objects->dynobj_end(); ++p) { Arm_dynobj* arm_dynobj = Arm_dynobj::as_arm_dynobj(*p); this->merge_processor_specific_flags( arm_dynobj->name(), arm_dynobj->processor_specific_flags()); this->merge_object_attributes(arm_dynobj->name().c_str(), arm_dynobj->attributes_section_data()); } // Check BLX use. const Object_attribute* cpu_arch_attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch); if (cpu_arch_attr->int_value() > elfcpp::TAG_CPU_ARCH_V4) this->set_may_use_blx(true); // Check if we need to use Cortex-A8 workaround. if (parameters->options().user_set_fix_cortex_a8()) this->fix_cortex_a8_ = parameters->options().fix_cortex_a8(); else { // If neither --fix-cortex-a8 nor --no-fix-cortex-a8 is used, turn on // Cortex-A8 erratum workaround for ARMv7-A or ARMv7 with unknown // profile. const Object_attribute* cpu_arch_profile_attr = this->get_aeabi_object_attribute(elfcpp::Tag_CPU_arch_profile); this->fix_cortex_a8_ = (cpu_arch_attr->int_value() == elfcpp::TAG_CPU_ARCH_V7 && (cpu_arch_profile_attr->int_value() == 'A' || cpu_arch_profile_attr->int_value() == 0)); } // Check if we can use V4BX interworking. // The V4BX interworking stub contains BX instruction, // which is not specified for some profiles. if (this->fix_v4bx() == General_options::FIX_V4BX_INTERWORKING && !this->may_use_blx()) gold_error(_("unable to provide V4BX reloc interworking fix up; " "the target profile does not support BX instruction")); // Fill in some more dynamic tags. const Reloc_section* rel_plt = (this->plt_ == NULL ? NULL : this->plt_->rel_plt()); layout->add_target_dynamic_tags(true, this->got_plt_, rel_plt, this->rel_dyn_, true, false); // Emit any relocs we saved in an attempt to avoid generating COPY // relocs. if (this->copy_relocs_.any_saved_relocs()) this->copy_relocs_.emit(this->rel_dyn_section(layout)); // Handle the .ARM.exidx section. Output_section* exidx_section = layout->find_output_section(".ARM.exidx"); if (exidx_section != NULL && exidx_section->type() == elfcpp::SHT_ARM_EXIDX && !parameters->options().relocatable()) { // Create __exidx_start and __exdix_end symbols. symtab->define_in_output_data("__exidx_start", NULL, Symbol_table::PREDEFINED, exidx_section, 0, 0, elfcpp::STT_OBJECT, elfcpp::STB_GLOBAL, elfcpp::STV_HIDDEN, 0, false, true); symtab->define_in_output_data("__exidx_end", NULL, Symbol_table::PREDEFINED, exidx_section, 0, 0, elfcpp::STT_OBJECT, elfcpp::STB_GLOBAL, elfcpp::STV_HIDDEN, 0, true, true); // For the ARM target, we need to add a PT_ARM_EXIDX segment for // the .ARM.exidx section. if (!layout->script_options()->saw_phdrs_clause()) { gold_assert(layout->find_output_segment(elfcpp::PT_ARM_EXIDX, 0, 0) == NULL); Output_segment* exidx_segment = layout->make_output_segment(elfcpp::PT_ARM_EXIDX, elfcpp::PF_R); exidx_segment->add_output_section(exidx_section, elfcpp::PF_R, false); } } // Create an .ARM.attributes section if there is not one already. Output_attributes_section_data* attributes_section = new Output_attributes_section_data(*this->attributes_section_data_); layout->add_output_section_data(".ARM.attributes", elfcpp::SHT_ARM_ATTRIBUTES, 0, attributes_section, false, false, false, false); } // Return whether a direct absolute static relocation needs to be applied. // In cases where Scan::local() or Scan::global() has created // a dynamic relocation other than R_ARM_RELATIVE, the addend // of the relocation is carried in the data, and we must not // apply the static relocation. template inline bool Target_arm::Relocate::should_apply_static_reloc( const Sized_symbol<32>* gsym, int ref_flags, bool is_32bit, Output_section* output_section) { // If the output section is not allocated, then we didn't call // scan_relocs, we didn't create a dynamic reloc, and we must apply // the reloc here. if ((output_section->flags() & elfcpp::SHF_ALLOC) == 0) return true; // For local symbols, we will have created a non-RELATIVE dynamic // relocation only if (a) the output is position independent, // (b) the relocation is absolute (not pc- or segment-relative), and // (c) the relocation is not 32 bits wide. if (gsym == NULL) return !(parameters->options().output_is_position_independent() && (ref_flags & Symbol::ABSOLUTE_REF) && !is_32bit); // For global symbols, we use the same helper routines used in the // scan pass. If we did not create a dynamic relocation, or if we // created a RELATIVE dynamic relocation, we should apply the static // relocation. bool has_dyn = gsym->needs_dynamic_reloc(ref_flags); bool is_rel = (ref_flags & Symbol::ABSOLUTE_REF) && gsym->can_use_relative_reloc(ref_flags & Symbol::FUNCTION_CALL); return !has_dyn || is_rel; } // Perform a relocation. template inline bool Target_arm::Relocate::relocate( const Relocate_info<32, big_endian>* relinfo, Target_arm* target, Output_section *output_section, size_t relnum, const elfcpp::Rel<32, big_endian>& rel, unsigned int r_type, const Sized_symbol<32>* gsym, const Symbol_value<32>* psymval, unsigned char* view, Arm_address address, section_size_type /* view_size */ ) { typedef Arm_relocate_functions Arm_relocate_functions; r_type = get_real_reloc_type(r_type); const Arm_reloc_property* reloc_property = arm_reloc_property_table->get_implemented_static_reloc_property(r_type); if (reloc_property == NULL) { std::string reloc_name = arm_reloc_property_table->reloc_name_in_error_message(r_type); gold_error_at_location(relinfo, relnum, rel.get_r_offset(), _("cannot relocate %s in object file"), reloc_name.c_str()); return true; } const Arm_relobj* object = Arm_relobj::as_arm_relobj(relinfo->object); // If the final branch target of a relocation is THUMB instruction, this // is 1. Otherwise it is 0. Arm_address thumb_bit = 0; Symbol_value<32> symval; bool is_weakly_undefined_without_plt = false; if (relnum != Target_arm::fake_relnum_for_stubs) { if (gsym != NULL) { // This is a global symbol. Determine if we use PLT and if the // final target is THUMB. if (gsym->use_plt_offset(reloc_is_non_pic(r_type))) { // This uses a PLT, change the symbol value. symval.set_output_value(target->plt_section()->address() + gsym->plt_offset()); psymval = &symval; } else if (gsym->is_weak_undefined()) { // This is a weakly undefined symbol and we do not use PLT // for this relocation. A branch targeting this symbol will // be converted into an NOP. is_weakly_undefined_without_plt = true; } else { // Set thumb bit if symbol: // -Has type STT_ARM_TFUNC or // -Has type STT_FUNC, is defined and with LSB in value set. thumb_bit = (((gsym->type() == elfcpp::STT_ARM_TFUNC) || (gsym->type() == elfcpp::STT_FUNC && !gsym->is_undefined() && ((psymval->value(object, 0) & 1) != 0))) ? 1 : 0); } } else { // This is a local symbol. Determine if the final target is THUMB. // We saved this information when all the local symbols were read. elfcpp::Elf_types<32>::Elf_WXword r_info = rel.get_r_info(); unsigned int r_sym = elfcpp::elf_r_sym<32>(r_info); thumb_bit = object->local_symbol_is_thumb_function(r_sym) ? 1 : 0; } } else { // This is a fake relocation synthesized for a stub. It does not have // a real symbol. We just look at the LSB of the symbol value to // determine if the target is THUMB or not. thumb_bit = ((psymval->value(object, 0) & 1) != 0); } // Strip LSB if this points to a THUMB target. if (thumb_bit != 0 && reloc_property->uses_thumb_bit() && ((psymval->value(object, 0) & 1) != 0)) { Arm_address stripped_value = psymval->value(object, 0) & ~static_cast(1); symval.set_output_value(stripped_value); psymval = &symval; } // Get the GOT offset if needed. // The GOT pointer points to the end of the GOT section. // We need to subtract the size of the GOT section to get // the actual offset to use in the relocation. bool have_got_offset = false; unsigned int got_offset = 0; switch (r_type) { case elfcpp::R_ARM_GOT_BREL: case elfcpp::R_ARM_GOT_PREL: if (gsym != NULL) { gold_assert(gsym->has_got_offset(GOT_TYPE_STANDARD)); got_offset = (gsym->got_offset(GOT_TYPE_STANDARD) - target->got_size()); } else { unsigned int r_sym = elfcpp::elf_r_sym<32>(rel.get_r_info()); gold_assert(object->local_has_got_offset(r_sym, GOT_TYPE_STANDARD)); got_offset = (object->local_got_offset(r_sym, GOT_TYPE_STANDARD) - target->got_size()); } have_got_offset = true; break; default: break; } // To look up relocation stubs, we need to pass the symbol table index of // a local symbol. unsigned int r_sym = elfcpp::elf_r_sym<32>(rel.get_r_info()); // Get the addressing origin of the output segment defining the // symbol gsym if needed (AAELF 4.6.1.2 Relocation types). Arm_address sym_origin = 0; if (reloc_property->uses_symbol_base()) { if (r_type == elfcpp::R_ARM_BASE_ABS && gsym == NULL) // R_ARM_BASE_ABS with the NULL symbol will give the // absolute address of the GOT origin (GOT_ORG) (see ARM IHI // 0044C (AAELF): 4.6.1.8 Proxy generating relocations). sym_origin = target->got_plt_section()->address(); else if (gsym == NULL) sym_origin = 0; else if (gsym->source() == Symbol::IN_OUTPUT_SEGMENT) sym_origin = gsym->output_segment()->vaddr(); else if (gsym->source() == Symbol::IN_OUTPUT_DATA) sym_origin = gsym->output_data()->address(); // TODO: Assumes the segment base to be zero for the global symbols // till the proper support for the segment-base-relative addressing // will be implemented. This is consistent with GNU ld. } // For relative addressing relocation, find out the relative address base. Arm_address relative_address_base = 0; switch(reloc_property->relative_address_base()) { case Arm_reloc_property::RAB_NONE: break; case Arm_reloc_property::RAB_B_S: relative_address_base = sym_origin; break; case Arm_reloc_property::RAB_GOT_ORG: relative_address_base = target->got_plt_section()->address(); break; case Arm_reloc_property::RAB_P: relative_address_base = address; break; case Arm_reloc_property::RAB_Pa: relative_address_base = address & 0xfffffffcU; break; default: gold_unreachable(); } typename Arm_relocate_functions::Status reloc_status = Arm_relocate_functions::STATUS_OKAY; bool check_overflow = reloc_property->checks_overflow(); switch (r_type) { case elfcpp::R_ARM_NONE: break; case elfcpp::R_ARM_ABS8: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::abs8(view, object, psymval); break; case elfcpp::R_ARM_ABS12: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::abs12(view, object, psymval); break; case elfcpp::R_ARM_ABS16: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::abs16(view, object, psymval); break; case elfcpp::R_ARM_ABS32: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, true, output_section)) reloc_status = Arm_relocate_functions::abs32(view, object, psymval, thumb_bit); break; case elfcpp::R_ARM_ABS32_NOI: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, true, output_section)) // No thumb bit for this relocation: (S + A) reloc_status = Arm_relocate_functions::abs32(view, object, psymval, 0); break; case elfcpp::R_ARM_MOVW_ABS_NC: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::movw(view, object, psymval, 0, thumb_bit, check_overflow); break; case elfcpp::R_ARM_MOVT_ABS: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::movt(view, object, psymval, 0); break; case elfcpp::R_ARM_THM_MOVW_ABS_NC: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::thm_movw(view, object, psymval, 0, thumb_bit, false); break; case elfcpp::R_ARM_THM_MOVT_ABS: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::thm_movt(view, object, psymval, 0); break; case elfcpp::R_ARM_MOVW_PREL_NC: case elfcpp::R_ARM_MOVW_BREL_NC: case elfcpp::R_ARM_MOVW_BREL: reloc_status = Arm_relocate_functions::movw(view, object, psymval, relative_address_base, thumb_bit, check_overflow); break; case elfcpp::R_ARM_MOVT_PREL: case elfcpp::R_ARM_MOVT_BREL: reloc_status = Arm_relocate_functions::movt(view, object, psymval, relative_address_base); break; case elfcpp::R_ARM_THM_MOVW_PREL_NC: case elfcpp::R_ARM_THM_MOVW_BREL_NC: case elfcpp::R_ARM_THM_MOVW_BREL: reloc_status = Arm_relocate_functions::thm_movw(view, object, psymval, relative_address_base, thumb_bit, check_overflow); break; case elfcpp::R_ARM_THM_MOVT_PREL: case elfcpp::R_ARM_THM_MOVT_BREL: reloc_status = Arm_relocate_functions::thm_movt(view, object, psymval, relative_address_base); break; case elfcpp::R_ARM_REL32: reloc_status = Arm_relocate_functions::rel32(view, object, psymval, address, thumb_bit); break; case elfcpp::R_ARM_THM_ABS5: if (should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) reloc_status = Arm_relocate_functions::thm_abs5(view, object, psymval); break; // Thumb long branches. case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_THM_XPC22: case elfcpp::R_ARM_THM_JUMP24: reloc_status = Arm_relocate_functions::thumb_branch_common( r_type, relinfo, view, gsym, object, r_sym, psymval, address, thumb_bit, is_weakly_undefined_without_plt); break; case elfcpp::R_ARM_GOTOFF32: { Arm_address got_origin; got_origin = target->got_plt_section()->address(); reloc_status = Arm_relocate_functions::rel32(view, object, psymval, got_origin, thumb_bit); } break; case elfcpp::R_ARM_BASE_PREL: gold_assert(gsym != NULL); reloc_status = Arm_relocate_functions::base_prel(view, sym_origin, address); break; case elfcpp::R_ARM_BASE_ABS: { if (!should_apply_static_reloc(gsym, Symbol::ABSOLUTE_REF, false, output_section)) break; reloc_status = Arm_relocate_functions::base_abs(view, sym_origin); } break; case elfcpp::R_ARM_GOT_BREL: gold_assert(have_got_offset); reloc_status = Arm_relocate_functions::got_brel(view, got_offset); break; case elfcpp::R_ARM_GOT_PREL: gold_assert(have_got_offset); // Get the address origin for GOT PLT, which is allocated right // after the GOT section, to calculate an absolute address of // the symbol GOT entry (got_origin + got_offset). Arm_address got_origin; got_origin = target->got_plt_section()->address(); reloc_status = Arm_relocate_functions::got_prel(view, got_origin + got_offset, address); break; case elfcpp::R_ARM_PLT32: case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_XPC25: gold_assert(gsym == NULL || gsym->has_plt_offset() || gsym->final_value_is_known() || (gsym->is_defined() && !gsym->is_from_dynobj() && !gsym->is_preemptible())); reloc_status = Arm_relocate_functions::arm_branch_common( r_type, relinfo, view, gsym, object, r_sym, psymval, address, thumb_bit, is_weakly_undefined_without_plt); break; case elfcpp::R_ARM_THM_JUMP19: reloc_status = Arm_relocate_functions::thm_jump19(view, object, psymval, address, thumb_bit); break; case elfcpp::R_ARM_THM_JUMP6: reloc_status = Arm_relocate_functions::thm_jump6(view, object, psymval, address); break; case elfcpp::R_ARM_THM_JUMP8: reloc_status = Arm_relocate_functions::thm_jump8(view, object, psymval, address); break; case elfcpp::R_ARM_THM_JUMP11: reloc_status = Arm_relocate_functions::thm_jump11(view, object, psymval, address); break; case elfcpp::R_ARM_PREL31: reloc_status = Arm_relocate_functions::prel31(view, object, psymval, address, thumb_bit); break; case elfcpp::R_ARM_V4BX: if (target->fix_v4bx() > General_options::FIX_V4BX_NONE) { const bool is_v4bx_interworking = (target->fix_v4bx() == General_options::FIX_V4BX_INTERWORKING); reloc_status = Arm_relocate_functions::v4bx(relinfo, view, object, address, is_v4bx_interworking); } break; case elfcpp::R_ARM_THM_PC8: reloc_status = Arm_relocate_functions::thm_pc8(view, object, psymval, address); break; case elfcpp::R_ARM_THM_PC12: reloc_status = Arm_relocate_functions::thm_pc12(view, object, psymval, address); break; case elfcpp::R_ARM_THM_ALU_PREL_11_0: reloc_status = Arm_relocate_functions::thm_alu11(view, object, psymval, address, thumb_bit); break; case elfcpp::R_ARM_ALU_PC_G0_NC: case elfcpp::R_ARM_ALU_PC_G0: case elfcpp::R_ARM_ALU_PC_G1_NC: case elfcpp::R_ARM_ALU_PC_G1: case elfcpp::R_ARM_ALU_PC_G2: case elfcpp::R_ARM_ALU_SB_G0_NC: case elfcpp::R_ARM_ALU_SB_G0: case elfcpp::R_ARM_ALU_SB_G1_NC: case elfcpp::R_ARM_ALU_SB_G1: case elfcpp::R_ARM_ALU_SB_G2: reloc_status = Arm_relocate_functions::arm_grp_alu(view, object, psymval, reloc_property->group_index(), relative_address_base, thumb_bit, check_overflow); break; case elfcpp::R_ARM_LDR_PC_G0: case elfcpp::R_ARM_LDR_PC_G1: case elfcpp::R_ARM_LDR_PC_G2: case elfcpp::R_ARM_LDR_SB_G0: case elfcpp::R_ARM_LDR_SB_G1: case elfcpp::R_ARM_LDR_SB_G2: reloc_status = Arm_relocate_functions::arm_grp_ldr(view, object, psymval, reloc_property->group_index(), relative_address_base); break; case elfcpp::R_ARM_LDRS_PC_G0: case elfcpp::R_ARM_LDRS_PC_G1: case elfcpp::R_ARM_LDRS_PC_G2: case elfcpp::R_ARM_LDRS_SB_G0: case elfcpp::R_ARM_LDRS_SB_G1: case elfcpp::R_ARM_LDRS_SB_G2: reloc_status = Arm_relocate_functions::arm_grp_ldrs(view, object, psymval, reloc_property->group_index(), relative_address_base); break; case elfcpp::R_ARM_LDC_PC_G0: case elfcpp::R_ARM_LDC_PC_G1: case elfcpp::R_ARM_LDC_PC_G2: case elfcpp::R_ARM_LDC_SB_G0: case elfcpp::R_ARM_LDC_SB_G1: case elfcpp::R_ARM_LDC_SB_G2: reloc_status = Arm_relocate_functions::arm_grp_ldc(view, object, psymval, reloc_property->group_index(), relative_address_base); break; default: gold_unreachable(); } // Report any errors. switch (reloc_status) { case Arm_relocate_functions::STATUS_OKAY: break; case Arm_relocate_functions::STATUS_OVERFLOW: gold_error_at_location(relinfo, relnum, rel.get_r_offset(), _("relocation overflow in relocation %u"), r_type); break; case Arm_relocate_functions::STATUS_BAD_RELOC: gold_error_at_location( relinfo, relnum, rel.get_r_offset(), _("unexpected opcode while processing relocation %u"), r_type); break; default: gold_unreachable(); } return true; } // Relocate section data. template void Target_arm::relocate_section( const Relocate_info<32, big_endian>* relinfo, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, unsigned char* view, Arm_address address, section_size_type view_size, const Reloc_symbol_changes* reloc_symbol_changes) { typedef typename Target_arm::Relocate Arm_relocate; gold_assert(sh_type == elfcpp::SHT_REL); // See if we are relocating a relaxed input section. If so, the view // covers the whole output section and we need to adjust accordingly. if (needs_special_offset_handling) { const Output_relaxed_input_section* poris = output_section->find_relaxed_input_section(relinfo->object, relinfo->data_shndx); if (poris != NULL) { Arm_address section_address = poris->address(); section_size_type section_size = poris->data_size(); gold_assert((section_address >= address) && ((section_address + section_size) <= (address + view_size))); off_t offset = section_address - address; view += offset; address += offset; view_size = section_size; } } gold::relocate_section<32, big_endian, Target_arm, elfcpp::SHT_REL, Arm_relocate>( relinfo, this, prelocs, reloc_count, output_section, needs_special_offset_handling, view, address, view_size, reloc_symbol_changes); } // Return the size of a relocation while scanning during a relocatable // link. template unsigned int Target_arm::Relocatable_size_for_reloc::get_size_for_reloc( unsigned int r_type, Relobj* object) { r_type = get_real_reloc_type(r_type); const Arm_reloc_property* arp = arm_reloc_property_table->get_implemented_static_reloc_property(r_type); if (arp != NULL) return arp->size(); else { std::string reloc_name = arm_reloc_property_table->reloc_name_in_error_message(r_type); gold_error(_("%s: unexpected %s in object file"), object->name().c_str(), reloc_name.c_str()); return 0; } } // Scan the relocs during a relocatable link. template void Target_arm::scan_relocatable_relocs( Symbol_table* symtab, Layout* layout, Sized_relobj<32, big_endian>* object, unsigned int data_shndx, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, const unsigned char* plocal_symbols, Relocatable_relocs* rr) { gold_assert(sh_type == elfcpp::SHT_REL); typedef gold::Default_scan_relocatable_relocs Scan_relocatable_relocs; gold::scan_relocatable_relocs<32, big_endian, elfcpp::SHT_REL, Scan_relocatable_relocs>( symtab, layout, object, data_shndx, prelocs, reloc_count, output_section, needs_special_offset_handling, local_symbol_count, plocal_symbols, rr); } // Relocate a section during a relocatable link. template void Target_arm::relocate_for_relocatable( const Relocate_info<32, big_endian>* relinfo, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, off_t offset_in_output_section, const Relocatable_relocs* rr, unsigned char* view, Arm_address view_address, section_size_type view_size, unsigned char* reloc_view, section_size_type reloc_view_size) { gold_assert(sh_type == elfcpp::SHT_REL); gold::relocate_for_relocatable<32, big_endian, elfcpp::SHT_REL>( relinfo, prelocs, reloc_count, output_section, offset_in_output_section, rr, view, view_address, view_size, reloc_view, reloc_view_size); } // Return the value to use for a dynamic symbol which requires special // treatment. This is how we support equality comparisons of function // pointers across shared library boundaries, as described in the // processor specific ABI supplement. template uint64_t Target_arm::do_dynsym_value(const Symbol* gsym) const { gold_assert(gsym->is_from_dynobj() && gsym->has_plt_offset()); return this->plt_section()->address() + gsym->plt_offset(); } // Map platform-specific relocs to real relocs // template unsigned int Target_arm::get_real_reloc_type (unsigned int r_type) { switch (r_type) { case elfcpp::R_ARM_TARGET1: // This is either R_ARM_ABS32 or R_ARM_REL32; return elfcpp::R_ARM_ABS32; case elfcpp::R_ARM_TARGET2: // This can be any reloc type but ususally is R_ARM_GOT_PREL return elfcpp::R_ARM_GOT_PREL; default: return r_type; } } // Whether if two EABI versions V1 and V2 are compatible. template bool Target_arm::are_eabi_versions_compatible( elfcpp::Elf_Word v1, elfcpp::Elf_Word v2) { // v4 and v5 are the same spec before and after it was released, // so allow mixing them. if ((v1 == elfcpp::EF_ARM_EABI_VER4 && v2 == elfcpp::EF_ARM_EABI_VER5) || (v1 == elfcpp::EF_ARM_EABI_VER5 && v2 == elfcpp::EF_ARM_EABI_VER4)) return true; return v1 == v2; } // Combine FLAGS from an input object called NAME and the processor-specific // flags in the ELF header of the output. Much of this is adapted from the // processor-specific flags merging code in elf32_arm_merge_private_bfd_data // in bfd/elf32-arm.c. template void Target_arm::merge_processor_specific_flags( const std::string& name, elfcpp::Elf_Word flags) { if (this->are_processor_specific_flags_set()) { elfcpp::Elf_Word out_flags = this->processor_specific_flags(); // Nothing to merge if flags equal to those in output. if (flags == out_flags) return; // Complain about various flag mismatches. elfcpp::Elf_Word version1 = elfcpp::arm_eabi_version(flags); elfcpp::Elf_Word version2 = elfcpp::arm_eabi_version(out_flags); if (!this->are_eabi_versions_compatible(version1, version2)) gold_error(_("Source object %s has EABI version %d but output has " "EABI version %d."), name.c_str(), (flags & elfcpp::EF_ARM_EABIMASK) >> 24, (out_flags & elfcpp::EF_ARM_EABIMASK) >> 24); } else { // If the input is the default architecture and had the default // flags then do not bother setting the flags for the output // architecture, instead allow future merges to do this. If no // future merges ever set these flags then they will retain their // uninitialised values, which surprise surprise, correspond // to the default values. if (flags == 0) return; // This is the first time, just copy the flags. // We only copy the EABI version for now. this->set_processor_specific_flags(flags & elfcpp::EF_ARM_EABIMASK); } } // Adjust ELF file header. template void Target_arm::do_adjust_elf_header( unsigned char* view, int len) const { gold_assert(len == elfcpp::Elf_sizes<32>::ehdr_size); elfcpp::Ehdr<32, big_endian> ehdr(view); unsigned char e_ident[elfcpp::EI_NIDENT]; memcpy(e_ident, ehdr.get_e_ident(), elfcpp::EI_NIDENT); if (elfcpp::arm_eabi_version(this->processor_specific_flags()) == elfcpp::EF_ARM_EABI_UNKNOWN) e_ident[elfcpp::EI_OSABI] = elfcpp::ELFOSABI_ARM; else e_ident[elfcpp::EI_OSABI] = 0; e_ident[elfcpp::EI_ABIVERSION] = 0; // FIXME: Do EF_ARM_BE8 adjustment. elfcpp::Ehdr_write<32, big_endian> oehdr(view); oehdr.put_e_ident(e_ident); } // do_make_elf_object to override the same function in the base class. // We need to use a target-specific sub-class of Sized_relobj<32, big_endian> // to store ARM specific information. Hence we need to have our own // ELF object creation. template Object* Target_arm::do_make_elf_object( const std::string& name, Input_file* input_file, off_t offset, const elfcpp::Ehdr<32, big_endian>& ehdr) { int et = ehdr.get_e_type(); if (et == elfcpp::ET_REL) { Arm_relobj* obj = new Arm_relobj(name, input_file, offset, ehdr); obj->setup(); return obj; } else if (et == elfcpp::ET_DYN) { Sized_dynobj<32, big_endian>* obj = new Arm_dynobj(name, input_file, offset, ehdr); obj->setup(); return obj; } else { gold_error(_("%s: unsupported ELF file type %d"), name.c_str(), et); return NULL; } } // Read the architecture from the Tag_also_compatible_with attribute, if any. // Returns -1 if no architecture could be read. // This is adapted from get_secondary_compatible_arch() in bfd/elf32-arm.c. template int Target_arm::get_secondary_compatible_arch( const Attributes_section_data* pasd) { const Object_attribute *known_attributes = pasd->known_attributes(Object_attribute::OBJ_ATTR_PROC); // Note: the tag and its argument below are uleb128 values, though // currently-defined values fit in one byte for each. const std::string& sv = known_attributes[elfcpp::Tag_also_compatible_with].string_value(); if (sv.size() == 2 && sv.data()[0] == elfcpp::Tag_CPU_arch && (sv.data()[1] & 128) != 128) return sv.data()[1]; // This tag is "safely ignorable", so don't complain if it looks funny. return -1; } // Set, or unset, the architecture of the Tag_also_compatible_with attribute. // The tag is removed if ARCH is -1. // This is adapted from set_secondary_compatible_arch() in bfd/elf32-arm.c. template void Target_arm::set_secondary_compatible_arch( Attributes_section_data* pasd, int arch) { Object_attribute *known_attributes = pasd->known_attributes(Object_attribute::OBJ_ATTR_PROC); if (arch == -1) { known_attributes[elfcpp::Tag_also_compatible_with].set_string_value(""); return; } // Note: the tag and its argument below are uleb128 values, though // currently-defined values fit in one byte for each. char sv[3]; sv[0] = elfcpp::Tag_CPU_arch; gold_assert(arch != 0); sv[1] = arch; sv[2] = '\0'; known_attributes[elfcpp::Tag_also_compatible_with].set_string_value(sv); } // Combine two values for Tag_CPU_arch, taking secondary compatibility tags // into account. // This is adapted from tag_cpu_arch_combine() in bfd/elf32-arm.c. template int Target_arm::tag_cpu_arch_combine( const char* name, int oldtag, int* secondary_compat_out, int newtag, int secondary_compat) { #define T(X) elfcpp::TAG_CPU_ARCH_##X static const int v6t2[] = { T(V6T2), // PRE_V4. T(V6T2), // V4. T(V6T2), // V4T. T(V6T2), // V5T. T(V6T2), // V5TE. T(V6T2), // V5TEJ. T(V6T2), // V6. T(V7), // V6KZ. T(V6T2) // V6T2. }; static const int v6k[] = { T(V6K), // PRE_V4. T(V6K), // V4. T(V6K), // V4T. T(V6K), // V5T. T(V6K), // V5TE. T(V6K), // V5TEJ. T(V6K), // V6. T(V6KZ), // V6KZ. T(V7), // V6T2. T(V6K) // V6K. }; static const int v7[] = { T(V7), // PRE_V4. T(V7), // V4. T(V7), // V4T. T(V7), // V5T. T(V7), // V5TE. T(V7), // V5TEJ. T(V7), // V6. T(V7), // V6KZ. T(V7), // V6T2. T(V7), // V6K. T(V7) // V7. }; static const int v6_m[] = { -1, // PRE_V4. -1, // V4. T(V6K), // V4T. T(V6K), // V5T. T(V6K), // V5TE. T(V6K), // V5TEJ. T(V6K), // V6. T(V6KZ), // V6KZ. T(V7), // V6T2. T(V6K), // V6K. T(V7), // V7. T(V6_M) // V6_M. }; static const int v6s_m[] = { -1, // PRE_V4. -1, // V4. T(V6K), // V4T. T(V6K), // V5T. T(V6K), // V5TE. T(V6K), // V5TEJ. T(V6K), // V6. T(V6KZ), // V6KZ. T(V7), // V6T2. T(V6K), // V6K. T(V7), // V7. T(V6S_M), // V6_M. T(V6S_M) // V6S_M. }; static const int v7e_m[] = { -1, // PRE_V4. -1, // V4. T(V7E_M), // V4T. T(V7E_M), // V5T. T(V7E_M), // V5TE. T(V7E_M), // V5TEJ. T(V7E_M), // V6. T(V7E_M), // V6KZ. T(V7E_M), // V6T2. T(V7E_M), // V6K. T(V7E_M), // V7. T(V7E_M), // V6_M. T(V7E_M), // V6S_M. T(V7E_M) // V7E_M. }; static const int v4t_plus_v6_m[] = { -1, // PRE_V4. -1, // V4. T(V4T), // V4T. T(V5T), // V5T. T(V5TE), // V5TE. T(V5TEJ), // V5TEJ. T(V6), // V6. T(V6KZ), // V6KZ. T(V6T2), // V6T2. T(V6K), // V6K. T(V7), // V7. T(V6_M), // V6_M. T(V6S_M), // V6S_M. T(V7E_M), // V7E_M. T(V4T_PLUS_V6_M) // V4T plus V6_M. }; static const int *comb[] = { v6t2, v6k, v7, v6_m, v6s_m, v7e_m, // Pseudo-architecture. v4t_plus_v6_m }; // Check we've not got a higher architecture than we know about. if (oldtag >= elfcpp::MAX_TAG_CPU_ARCH || newtag >= elfcpp::MAX_TAG_CPU_ARCH) { gold_error(_("%s: unknown CPU architecture"), name); return -1; } // Override old tag if we have a Tag_also_compatible_with on the output. if ((oldtag == T(V6_M) && *secondary_compat_out == T(V4T)) || (oldtag == T(V4T) && *secondary_compat_out == T(V6_M))) oldtag = T(V4T_PLUS_V6_M); // And override the new tag if we have a Tag_also_compatible_with on the // input. if ((newtag == T(V6_M) && secondary_compat == T(V4T)) || (newtag == T(V4T) && secondary_compat == T(V6_M))) newtag = T(V4T_PLUS_V6_M); // Architectures before V6KZ add features monotonically. int tagh = std::max(oldtag, newtag); if (tagh <= elfcpp::TAG_CPU_ARCH_V6KZ) return tagh; int tagl = std::min(oldtag, newtag); int result = comb[tagh - T(V6T2)][tagl]; // Use Tag_CPU_arch == V4T and Tag_also_compatible_with (Tag_CPU_arch V6_M) // as the canonical version. if (result == T(V4T_PLUS_V6_M)) { result = T(V4T); *secondary_compat_out = T(V6_M); } else *secondary_compat_out = -1; if (result == -1) { gold_error(_("%s: conflicting CPU architectures %d/%d"), name, oldtag, newtag); return -1; } return result; #undef T } // Helper to print AEABI enum tag value. template std::string Target_arm::aeabi_enum_name(unsigned int value) { static const char *aeabi_enum_names[] = { "", "variable-size", "32-bit", "" }; const size_t aeabi_enum_names_size = sizeof(aeabi_enum_names) / sizeof(aeabi_enum_names[0]); if (value < aeabi_enum_names_size) return std::string(aeabi_enum_names[value]); else { char buffer[100]; sprintf(buffer, "", value); return std::string(buffer); } } // Return the string value to store in TAG_CPU_name. template std::string Target_arm::tag_cpu_name_value(unsigned int value) { static const char *name_table[] = { // These aren't real CPU names, but we can't guess // that from the architecture version alone. "Pre v4", "ARM v4", "ARM v4T", "ARM v5T", "ARM v5TE", "ARM v5TEJ", "ARM v6", "ARM v6KZ", "ARM v6T2", "ARM v6K", "ARM v7", "ARM v6-M", "ARM v6S-M", "ARM v7E-M" }; const size_t name_table_size = sizeof(name_table) / sizeof(name_table[0]); if (value < name_table_size) return std::string(name_table[value]); else { char buffer[100]; sprintf(buffer, "", value); return std::string(buffer); } } // Merge object attributes from input file called NAME with those of the // output. The input object attributes are in the object pointed by PASD. template void Target_arm::merge_object_attributes( const char* name, const Attributes_section_data* pasd) { // Return if there is no attributes section data. if (pasd == NULL) return; // If output has no object attributes, just copy. if (this->attributes_section_data_ == NULL) { this->attributes_section_data_ = new Attributes_section_data(*pasd); return; } const int vendor = Object_attribute::OBJ_ATTR_PROC; const Object_attribute* in_attr = pasd->known_attributes(vendor); Object_attribute* out_attr = this->attributes_section_data_->known_attributes(vendor); // This needs to happen before Tag_ABI_FP_number_model is merged. */ if (in_attr[elfcpp::Tag_ABI_VFP_args].int_value() != out_attr[elfcpp::Tag_ABI_VFP_args].int_value()) { // Ignore mismatches if the object doesn't use floating point. */ if (out_attr[elfcpp::Tag_ABI_FP_number_model].int_value() == 0) out_attr[elfcpp::Tag_ABI_VFP_args].set_int_value( in_attr[elfcpp::Tag_ABI_VFP_args].int_value()); else if (in_attr[elfcpp::Tag_ABI_FP_number_model].int_value() != 0) gold_error(_("%s uses VFP register arguments, output does not"), name); } for (int i = 4; i < Vendor_object_attributes::NUM_KNOWN_ATTRIBUTES; ++i) { // Merge this attribute with existing attributes. switch (i) { case elfcpp::Tag_CPU_raw_name: case elfcpp::Tag_CPU_name: // These are merged after Tag_CPU_arch. break; case elfcpp::Tag_ABI_optimization_goals: case elfcpp::Tag_ABI_FP_optimization_goals: // Use the first value seen. break; case elfcpp::Tag_CPU_arch: { unsigned int saved_out_attr = out_attr->int_value(); // Merge Tag_CPU_arch and Tag_also_compatible_with. int secondary_compat = this->get_secondary_compatible_arch(pasd); int secondary_compat_out = this->get_secondary_compatible_arch( this->attributes_section_data_); out_attr[i].set_int_value( tag_cpu_arch_combine(name, out_attr[i].int_value(), &secondary_compat_out, in_attr[i].int_value(), secondary_compat)); this->set_secondary_compatible_arch(this->attributes_section_data_, secondary_compat_out); // Merge Tag_CPU_name and Tag_CPU_raw_name. if (out_attr[i].int_value() == saved_out_attr) ; // Leave the names alone. else if (out_attr[i].int_value() == in_attr[i].int_value()) { // The output architecture has been changed to match the // input architecture. Use the input names. out_attr[elfcpp::Tag_CPU_name].set_string_value( in_attr[elfcpp::Tag_CPU_name].string_value()); out_attr[elfcpp::Tag_CPU_raw_name].set_string_value( in_attr[elfcpp::Tag_CPU_raw_name].string_value()); } else { out_attr[elfcpp::Tag_CPU_name].set_string_value(""); out_attr[elfcpp::Tag_CPU_raw_name].set_string_value(""); } // If we still don't have a value for Tag_CPU_name, // make one up now. Tag_CPU_raw_name remains blank. if (out_attr[elfcpp::Tag_CPU_name].string_value() == "") { const std::string cpu_name = this->tag_cpu_name_value(out_attr[i].int_value()); // FIXME: If we see an unknown CPU, this will be set // to "", where n is the attribute value. // This is different from BFD, which leaves the name alone. out_attr[elfcpp::Tag_CPU_name].set_string_value(cpu_name); } } break; case elfcpp::Tag_ARM_ISA_use: case elfcpp::Tag_THUMB_ISA_use: case elfcpp::Tag_WMMX_arch: case elfcpp::Tag_Advanced_SIMD_arch: // ??? Do Advanced_SIMD (NEON) and WMMX conflict? case elfcpp::Tag_ABI_FP_rounding: case elfcpp::Tag_ABI_FP_exceptions: case elfcpp::Tag_ABI_FP_user_exceptions: case elfcpp::Tag_ABI_FP_number_model: case elfcpp::Tag_VFP_HP_extension: case elfcpp::Tag_CPU_unaligned_access: case elfcpp::Tag_T2EE_use: case elfcpp::Tag_Virtualization_use: case elfcpp::Tag_MPextension_use: // Use the largest value specified. if (in_attr[i].int_value() > out_attr[i].int_value()) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_align8_preserved: case elfcpp::Tag_ABI_PCS_RO_data: // Use the smallest value specified. if (in_attr[i].int_value() < out_attr[i].int_value()) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_align8_needed: if ((in_attr[i].int_value() > 0 || out_attr[i].int_value() > 0) && (in_attr[elfcpp::Tag_ABI_align8_preserved].int_value() == 0 || (out_attr[elfcpp::Tag_ABI_align8_preserved].int_value() == 0))) { // This error message should be enabled once all non-conformant // binaries in the toolchain have had the attributes set // properly. // gold_error(_("output 8-byte data alignment conflicts with %s"), // name); } // Fall through. case elfcpp::Tag_ABI_FP_denormal: case elfcpp::Tag_ABI_PCS_GOT_use: { // These tags have 0 = don't care, 1 = strong requirement, // 2 = weak requirement. static const int order_021[3] = {0, 2, 1}; // Use the "greatest" from the sequence 0, 2, 1, or the largest // value if greater than 2 (for future-proofing). if ((in_attr[i].int_value() > 2 && in_attr[i].int_value() > out_attr[i].int_value()) || (in_attr[i].int_value() <= 2 && out_attr[i].int_value() <= 2 && (order_021[in_attr[i].int_value()] > order_021[out_attr[i].int_value()]))) out_attr[i].set_int_value(in_attr[i].int_value()); } break; case elfcpp::Tag_CPU_arch_profile: if (out_attr[i].int_value() != in_attr[i].int_value()) { // 0 will merge with anything. // 'A' and 'S' merge to 'A'. // 'R' and 'S' merge to 'R'. // 'M' and 'A|R|S' is an error. if (out_attr[i].int_value() == 0 || (out_attr[i].int_value() == 'S' && (in_attr[i].int_value() == 'A' || in_attr[i].int_value() == 'R'))) out_attr[i].set_int_value(in_attr[i].int_value()); else if (in_attr[i].int_value() == 0 || (in_attr[i].int_value() == 'S' && (out_attr[i].int_value() == 'A' || out_attr[i].int_value() == 'R'))) ; // Do nothing. else { gold_error (_("conflicting architecture profiles %c/%c"), in_attr[i].int_value() ? in_attr[i].int_value() : '0', out_attr[i].int_value() ? out_attr[i].int_value() : '0'); } } break; case elfcpp::Tag_VFP_arch: { static const struct { int ver; int regs; } vfp_versions[7] = { {0, 0}, {1, 16}, {2, 16}, {3, 32}, {3, 16}, {4, 32}, {4, 16} }; // Values greater than 6 aren't defined, so just pick the // biggest. if (in_attr[i].int_value() > 6 && in_attr[i].int_value() > out_attr[i].int_value()) { *out_attr = *in_attr; break; } // The output uses the superset of input features // (ISA version) and registers. int ver = std::max(vfp_versions[in_attr[i].int_value()].ver, vfp_versions[out_attr[i].int_value()].ver); int regs = std::max(vfp_versions[in_attr[i].int_value()].regs, vfp_versions[out_attr[i].int_value()].regs); // This assumes all possible supersets are also a valid // options. int newval; for (newval = 6; newval > 0; newval--) { if (regs == vfp_versions[newval].regs && ver == vfp_versions[newval].ver) break; } out_attr[i].set_int_value(newval); } break; case elfcpp::Tag_PCS_config: if (out_attr[i].int_value() == 0) out_attr[i].set_int_value(in_attr[i].int_value()); else if (in_attr[i].int_value() != 0 && out_attr[i].int_value() != 0) { // It's sometimes ok to mix different configs, so this is only // a warning. gold_warning(_("%s: conflicting platform configuration"), name); } break; case elfcpp::Tag_ABI_PCS_R9_use: if (in_attr[i].int_value() != out_attr[i].int_value() && out_attr[i].int_value() != elfcpp::AEABI_R9_unused && in_attr[i].int_value() != elfcpp::AEABI_R9_unused) { gold_error(_("%s: conflicting use of R9"), name); } if (out_attr[i].int_value() == elfcpp::AEABI_R9_unused) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_PCS_RW_data: if (in_attr[i].int_value() == elfcpp::AEABI_PCS_RW_data_SBrel && (in_attr[elfcpp::Tag_ABI_PCS_R9_use].int_value() != elfcpp::AEABI_R9_SB) && (out_attr[elfcpp::Tag_ABI_PCS_R9_use].int_value() != elfcpp::AEABI_R9_unused)) { gold_error(_("%s: SB relative addressing conflicts with use " "of R9"), name); } // Use the smallest value specified. if (in_attr[i].int_value() < out_attr[i].int_value()) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_PCS_wchar_t: // FIXME: Make it possible to turn off this warning. if (out_attr[i].int_value() && in_attr[i].int_value() && out_attr[i].int_value() != in_attr[i].int_value()) { gold_warning(_("%s uses %u-byte wchar_t yet the output is to " "use %u-byte wchar_t; use of wchar_t values " "across objects may fail"), name, in_attr[i].int_value(), out_attr[i].int_value()); } else if (in_attr[i].int_value() && !out_attr[i].int_value()) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_enum_size: if (in_attr[i].int_value() != elfcpp::AEABI_enum_unused) { if (out_attr[i].int_value() == elfcpp::AEABI_enum_unused || out_attr[i].int_value() == elfcpp::AEABI_enum_forced_wide) { // The existing object is compatible with anything. // Use whatever requirements the new object has. out_attr[i].set_int_value(in_attr[i].int_value()); } // FIXME: Make it possible to turn off this warning. else if (in_attr[i].int_value() != elfcpp::AEABI_enum_forced_wide && out_attr[i].int_value() != in_attr[i].int_value()) { unsigned int in_value = in_attr[i].int_value(); unsigned int out_value = out_attr[i].int_value(); gold_warning(_("%s uses %s enums yet the output is to use " "%s enums; use of enum values across objects " "may fail"), name, this->aeabi_enum_name(in_value).c_str(), this->aeabi_enum_name(out_value).c_str()); } } break; case elfcpp::Tag_ABI_VFP_args: // Aready done. break; case elfcpp::Tag_ABI_WMMX_args: if (in_attr[i].int_value() != out_attr[i].int_value()) { gold_error(_("%s uses iWMMXt register arguments, output does " "not"), name); } break; case Object_attribute::Tag_compatibility: // Merged in target-independent code. break; case elfcpp::Tag_ABI_HardFP_use: // 1 (SP) and 2 (DP) conflict, so combine to 3 (SP & DP). if ((in_attr[i].int_value() == 1 && out_attr[i].int_value() == 2) || (in_attr[i].int_value() == 2 && out_attr[i].int_value() == 1)) out_attr[i].set_int_value(3); else if (in_attr[i].int_value() > out_attr[i].int_value()) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_ABI_FP_16bit_format: if (in_attr[i].int_value() != 0 && out_attr[i].int_value() != 0) { if (in_attr[i].int_value() != out_attr[i].int_value()) gold_error(_("fp16 format mismatch between %s and output"), name); } if (in_attr[i].int_value() != 0) out_attr[i].set_int_value(in_attr[i].int_value()); break; case elfcpp::Tag_nodefaults: // This tag is set if it exists, but the value is unused (and is // typically zero). We don't actually need to do anything here - // the merge happens automatically when the type flags are merged // below. break; case elfcpp::Tag_also_compatible_with: // Already done in Tag_CPU_arch. break; case elfcpp::Tag_conformance: // Keep the attribute if it matches. Throw it away otherwise. // No attribute means no claim to conform. if (in_attr[i].string_value() != out_attr[i].string_value()) out_attr[i].set_string_value(""); break; default: { const char* err_object = NULL; // The "known_obj_attributes" table does contain some undefined // attributes. Ensure that there are unused. if (out_attr[i].int_value() != 0 || out_attr[i].string_value() != "") err_object = "output"; else if (in_attr[i].int_value() != 0 || in_attr[i].string_value() != "") err_object = name; if (err_object != NULL) { // Attribute numbers >=64 (mod 128) can be safely ignored. if ((i & 127) < 64) gold_error(_("%s: unknown mandatory EABI object attribute " "%d"), err_object, i); else gold_warning(_("%s: unknown EABI object attribute %d"), err_object, i); } // Only pass on attributes that match in both inputs. if (!in_attr[i].matches(out_attr[i])) { out_attr[i].set_int_value(0); out_attr[i].set_string_value(""); } } } // If out_attr was copied from in_attr then it won't have a type yet. if (in_attr[i].type() && !out_attr[i].type()) out_attr[i].set_type(in_attr[i].type()); } // Merge Tag_compatibility attributes and any common GNU ones. this->attributes_section_data_->merge(name, pasd); // Check for any attributes not known on ARM. typedef Vendor_object_attributes::Other_attributes Other_attributes; const Other_attributes* in_other_attributes = pasd->other_attributes(vendor); Other_attributes::const_iterator in_iter = in_other_attributes->begin(); Other_attributes* out_other_attributes = this->attributes_section_data_->other_attributes(vendor); Other_attributes::iterator out_iter = out_other_attributes->begin(); while (in_iter != in_other_attributes->end() || out_iter != out_other_attributes->end()) { const char* err_object = NULL; int err_tag = 0; // The tags for each list are in numerical order. // If the tags are equal, then merge. if (out_iter != out_other_attributes->end() && (in_iter == in_other_attributes->end() || in_iter->first > out_iter->first)) { // This attribute only exists in output. We can't merge, and we // don't know what the tag means, so delete it. err_object = "output"; err_tag = out_iter->first; int saved_tag = out_iter->first; delete out_iter->second; out_other_attributes->erase(out_iter); out_iter = out_other_attributes->upper_bound(saved_tag); } else if (in_iter != in_other_attributes->end() && (out_iter != out_other_attributes->end() || in_iter->first < out_iter->first)) { // This attribute only exists in input. We can't merge, and we // don't know what the tag means, so ignore it. err_object = name; err_tag = in_iter->first; ++in_iter; } else // The tags are equal. { // As present, all attributes in the list are unknown, and // therefore can't be merged meaningfully. err_object = "output"; err_tag = out_iter->first; // Only pass on attributes that match in both inputs. if (!in_iter->second->matches(*(out_iter->second))) { // No match. Delete the attribute. int saved_tag = out_iter->first; delete out_iter->second; out_other_attributes->erase(out_iter); out_iter = out_other_attributes->upper_bound(saved_tag); } else { // Matched. Keep the attribute and move to the next. ++out_iter; ++in_iter; } } if (err_object) { // Attribute numbers >=64 (mod 128) can be safely ignored. */ if ((err_tag & 127) < 64) { gold_error(_("%s: unknown mandatory EABI object attribute %d"), err_object, err_tag); } else { gold_warning(_("%s: unknown EABI object attribute %d"), err_object, err_tag); } } } } // Stub-generation methods for Target_arm. // Make a new Arm_input_section object. template Arm_input_section* Target_arm::new_arm_input_section( Relobj* relobj, unsigned int shndx) { Section_id sid(relobj, shndx); Arm_input_section* arm_input_section = new Arm_input_section(relobj, shndx); arm_input_section->init(); // Register new Arm_input_section in map for look-up. std::pair ins = this->arm_input_section_map_.insert(std::make_pair(sid, arm_input_section)); // Make sure that it we have not created another Arm_input_section // for this input section already. gold_assert(ins.second); return arm_input_section; } // Find the Arm_input_section object corresponding to the SHNDX-th input // section of RELOBJ. template Arm_input_section* Target_arm::find_arm_input_section( Relobj* relobj, unsigned int shndx) const { Section_id sid(relobj, shndx); typename Arm_input_section_map::const_iterator p = this->arm_input_section_map_.find(sid); return (p != this->arm_input_section_map_.end()) ? p->second : NULL; } // Make a new stub table. template Stub_table* Target_arm::new_stub_table(Arm_input_section* owner) { Stub_table* stub_table = new Stub_table(owner); this->stub_tables_.push_back(stub_table); stub_table->set_address(owner->address() + owner->data_size()); stub_table->set_file_offset(owner->offset() + owner->data_size()); stub_table->finalize_data_size(); return stub_table; } // Scan a relocation for stub generation. template void Target_arm::scan_reloc_for_stub( const Relocate_info<32, big_endian>* relinfo, unsigned int r_type, const Sized_symbol<32>* gsym, unsigned int r_sym, const Symbol_value<32>* psymval, elfcpp::Elf_types<32>::Elf_Swxword addend, Arm_address address) { typedef typename Target_arm::Relocate Relocate; const Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(relinfo->object); if (r_type == elfcpp::R_ARM_V4BX) { const uint32_t reg = (addend & 0xf); if (this->fix_v4bx() == General_options::FIX_V4BX_INTERWORKING && reg < 0xf) { // Try looking up an existing stub from a stub table. Stub_table* stub_table = arm_relobj->stub_table(relinfo->data_shndx); gold_assert(stub_table != NULL); if (stub_table->find_arm_v4bx_stub(reg) == NULL) { // create a new stub and add it to stub table. Arm_v4bx_stub* stub = this->stub_factory().make_arm_v4bx_stub(reg); gold_assert(stub != NULL); stub_table->add_arm_v4bx_stub(stub); } } return; } bool target_is_thumb; Symbol_value<32> symval; if (gsym != NULL) { // This is a global symbol. Determine if we use PLT and if the // final target is THUMB. if (gsym->use_plt_offset(Relocate::reloc_is_non_pic(r_type))) { // This uses a PLT, change the symbol value. symval.set_output_value(this->plt_section()->address() + gsym->plt_offset()); psymval = &symval; target_is_thumb = false; } else if (gsym->is_undefined()) // There is no need to generate a stub symbol is undefined. return; else { target_is_thumb = ((gsym->type() == elfcpp::STT_ARM_TFUNC) || (gsym->type() == elfcpp::STT_FUNC && !gsym->is_undefined() && ((psymval->value(arm_relobj, 0) & 1) != 0))); } } else { // This is a local symbol. Determine if the final target is THUMB. target_is_thumb = arm_relobj->local_symbol_is_thumb_function(r_sym); } // Strip LSB if this points to a THUMB target. const Arm_reloc_property* reloc_property = arm_reloc_property_table->get_implemented_static_reloc_property(r_type); gold_assert(reloc_property != NULL); if (target_is_thumb && reloc_property->uses_thumb_bit() && ((psymval->value(arm_relobj, 0) & 1) != 0)) { Arm_address stripped_value = psymval->value(arm_relobj, 0) & ~static_cast(1); symval.set_output_value(stripped_value); psymval = &symval; } // Get the symbol value. Symbol_value<32>::Value value = psymval->value(arm_relobj, 0); // Owing to pipelining, the PC relative branches below actually skip // two instructions when the branch offset is 0. Arm_address destination; switch (r_type) { case elfcpp::R_ARM_CALL: case elfcpp::R_ARM_JUMP24: case elfcpp::R_ARM_PLT32: // ARM branches. destination = value + addend + 8; break; case elfcpp::R_ARM_THM_CALL: case elfcpp::R_ARM_THM_XPC22: case elfcpp::R_ARM_THM_JUMP24: case elfcpp::R_ARM_THM_JUMP19: // THUMB branches. destination = value + addend + 4; break; default: gold_unreachable(); } Reloc_stub* stub = NULL; Stub_type stub_type = Reloc_stub::stub_type_for_reloc(r_type, address, destination, target_is_thumb); if (stub_type != arm_stub_none) { // Try looking up an existing stub from a stub table. Stub_table* stub_table = arm_relobj->stub_table(relinfo->data_shndx); gold_assert(stub_table != NULL); // Locate stub by destination. Reloc_stub::Key stub_key(stub_type, gsym, arm_relobj, r_sym, addend); // Create a stub if there is not one already stub = stub_table->find_reloc_stub(stub_key); if (stub == NULL) { // create a new stub and add it to stub table. stub = this->stub_factory().make_reloc_stub(stub_type); stub_table->add_reloc_stub(stub, stub_key); } // Record the destination address. stub->set_destination_address(destination | (target_is_thumb ? 1 : 0)); } // For Cortex-A8, we need to record a relocation at 4K page boundary. if (this->fix_cortex_a8_ && (r_type == elfcpp::R_ARM_THM_JUMP24 || r_type == elfcpp::R_ARM_THM_JUMP19 || r_type == elfcpp::R_ARM_THM_CALL || r_type == elfcpp::R_ARM_THM_XPC22) && (address & 0xfffU) == 0xffeU) { // Found a candidate. Note we haven't checked the destination is // within 4K here: if we do so (and don't create a record) we can't // tell that a branch should have been relocated when scanning later. this->cortex_a8_relocs_info_[address] = new Cortex_a8_reloc(stub, r_type, destination | (target_is_thumb ? 1 : 0)); } } // This function scans a relocation sections for stub generation. // The template parameter Relocate must be a class type which provides // a single function, relocate(), which implements the machine // specific part of a relocation. // BIG_ENDIAN is the endianness of the data. SH_TYPE is the section type: // SHT_REL or SHT_RELA. // PRELOCS points to the relocation data. RELOC_COUNT is the number // of relocs. OUTPUT_SECTION is the output section. // NEEDS_SPECIAL_OFFSET_HANDLING is true if input offsets need to be // mapped to output offsets. // VIEW is the section data, VIEW_ADDRESS is its memory address, and // VIEW_SIZE is the size. These refer to the input section, unless // NEEDS_SPECIAL_OFFSET_HANDLING is true, in which case they refer to // the output section. template template void inline Target_arm::scan_reloc_section_for_stubs( const Relocate_info<32, big_endian>* relinfo, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, const unsigned char* view, elfcpp::Elf_types<32>::Elf_Addr view_address, section_size_type) { typedef typename Reloc_types::Reloc Reltype; const int reloc_size = Reloc_types::reloc_size; Arm_relobj* arm_object = Arm_relobj::as_arm_relobj(relinfo->object); unsigned int local_count = arm_object->local_symbol_count(); Comdat_behavior comdat_behavior = CB_UNDETERMINED; for (size_t i = 0; i < reloc_count; ++i, prelocs += reloc_size) { Reltype reloc(prelocs); typename elfcpp::Elf_types<32>::Elf_WXword r_info = reloc.get_r_info(); unsigned int r_sym = elfcpp::elf_r_sym<32>(r_info); unsigned int r_type = elfcpp::elf_r_type<32>(r_info); r_type = this->get_real_reloc_type(r_type); // Only a few relocation types need stubs. if ((r_type != elfcpp::R_ARM_CALL) && (r_type != elfcpp::R_ARM_JUMP24) && (r_type != elfcpp::R_ARM_PLT32) && (r_type != elfcpp::R_ARM_THM_CALL) && (r_type != elfcpp::R_ARM_THM_XPC22) && (r_type != elfcpp::R_ARM_THM_JUMP24) && (r_type != elfcpp::R_ARM_THM_JUMP19) && (r_type != elfcpp::R_ARM_V4BX)) continue; section_offset_type offset = convert_to_section_size_type(reloc.get_r_offset()); if (needs_special_offset_handling) { offset = output_section->output_offset(relinfo->object, relinfo->data_shndx, offset); if (offset == -1) continue; } if (r_type == elfcpp::R_ARM_V4BX) { // Get the BX instruction. typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(view + offset); elfcpp::Elf_types<32>::Elf_Swxword insn = elfcpp::Swap<32, big_endian>::readval(wv); this->scan_reloc_for_stub(relinfo, r_type, NULL, 0, NULL, insn, NULL); continue; } // Get the addend. Stub_addend_reader stub_addend_reader; elfcpp::Elf_types<32>::Elf_Swxword addend = stub_addend_reader(r_type, view + offset, reloc); const Sized_symbol<32>* sym; Symbol_value<32> symval; const Symbol_value<32> *psymval; if (r_sym < local_count) { sym = NULL; psymval = arm_object->local_symbol(r_sym); // If the local symbol belongs to a section we are discarding, // and that section is a debug section, try to find the // corresponding kept section and map this symbol to its // counterpart in the kept section. The symbol must not // correspond to a section we are folding. bool is_ordinary; unsigned int shndx = psymval->input_shndx(&is_ordinary); if (is_ordinary && shndx != elfcpp::SHN_UNDEF && !arm_object->is_section_included(shndx) && !(relinfo->symtab->is_section_folded(arm_object, shndx))) { if (comdat_behavior == CB_UNDETERMINED) { std::string name = arm_object->section_name(relinfo->data_shndx); comdat_behavior = get_comdat_behavior(name.c_str()); } if (comdat_behavior == CB_PRETEND) { bool found; typename elfcpp::Elf_types<32>::Elf_Addr value = arm_object->map_to_kept_section(shndx, &found); if (found) symval.set_output_value(value + psymval->input_value()); else symval.set_output_value(0); } else { symval.set_output_value(0); } symval.set_no_output_symtab_entry(); psymval = &symval; } } else { const Symbol* gsym = arm_object->global_symbol(r_sym); gold_assert(gsym != NULL); if (gsym->is_forwarder()) gsym = relinfo->symtab->resolve_forwards(gsym); sym = static_cast*>(gsym); if (sym->has_symtab_index()) symval.set_output_symtab_index(sym->symtab_index()); else symval.set_no_output_symtab_entry(); // We need to compute the would-be final value of this global // symbol. const Symbol_table* symtab = relinfo->symtab; const Sized_symbol<32>* sized_symbol = symtab->get_sized_symbol<32>(gsym); Symbol_table::Compute_final_value_status status; Arm_address value = symtab->compute_final_value<32>(sized_symbol, &status); // Skip this if the symbol has not output section. if (status == Symbol_table::CFVS_NO_OUTPUT_SECTION) continue; symval.set_output_value(value); psymval = &symval; } // If symbol is a section symbol, we don't know the actual type of // destination. Give up. if (psymval->is_section_symbol()) continue; this->scan_reloc_for_stub(relinfo, r_type, sym, r_sym, psymval, addend, view_address + offset); } } // Scan an input section for stub generation. template void Target_arm::scan_section_for_stubs( const Relocate_info<32, big_endian>* relinfo, unsigned int sh_type, const unsigned char* prelocs, size_t reloc_count, Output_section* output_section, bool needs_special_offset_handling, const unsigned char* view, Arm_address view_address, section_size_type view_size) { if (sh_type == elfcpp::SHT_REL) this->scan_reloc_section_for_stubs( relinfo, prelocs, reloc_count, output_section, needs_special_offset_handling, view, view_address, view_size); else if (sh_type == elfcpp::SHT_RELA) // We do not support RELA type relocations yet. This is provided for // completeness. this->scan_reloc_section_for_stubs( relinfo, prelocs, reloc_count, output_section, needs_special_offset_handling, view, view_address, view_size); else gold_unreachable(); } // Group input sections for stub generation. // // We goup input sections in an output sections so that the total size, // including any padding space due to alignment is smaller than GROUP_SIZE // unless the only input section in group is bigger than GROUP_SIZE already. // Then an ARM stub table is created to follow the last input section // in group. For each group an ARM stub table is created an is placed // after the last group. If STUB_ALWATS_AFTER_BRANCH is false, we further // extend the group after the stub table. template void Target_arm::group_sections( Layout* layout, section_size_type group_size, bool stubs_always_after_branch) { // Group input sections and insert stub table Layout::Section_list section_list; layout->get_allocated_sections(§ion_list); for (Layout::Section_list::const_iterator p = section_list.begin(); p != section_list.end(); ++p) { Arm_output_section* output_section = Arm_output_section::as_arm_output_section(*p); output_section->group_sections(group_size, stubs_always_after_branch, this); } } // Relaxation hook. This is where we do stub generation. template bool Target_arm::do_relax( int pass, const Input_objects* input_objects, Symbol_table* symtab, Layout* layout) { // No need to generate stubs if this is a relocatable link. gold_assert(!parameters->options().relocatable()); // If this is the first pass, we need to group input sections into // stub groups. bool done_exidx_fixup = false; if (pass == 1) { // Determine the stub group size. The group size is the absolute // value of the parameter --stub-group-size. If --stub-group-size // is passed a negative value, we restict stubs to be always after // the stubbed branches. int32_t stub_group_size_param = parameters->options().stub_group_size(); bool stubs_always_after_branch = stub_group_size_param < 0; section_size_type stub_group_size = abs(stub_group_size_param); // The Cortex-A8 erratum fix depends on stubs not being in the same 4K // page as the first half of a 32-bit branch straddling two 4K pages. // This is a crude way of enforcing that. if (this->fix_cortex_a8_) stubs_always_after_branch = true; if (stub_group_size == 1) { // Default value. // Thumb branch range is +-4MB has to be used as the default // maximum size (a given section can contain both ARM and Thumb // code, so the worst case has to be taken into account). // // This value is 24K less than that, which allows for 2025 // 12-byte stubs. If we exceed that, then we will fail to link. // The user will have to relink with an explicit group size // option. stub_group_size = 4170000; } group_sections(layout, stub_group_size, stubs_always_after_branch); // Also fix .ARM.exidx section coverage. Output_section* os = layout->find_output_section(".ARM.exidx"); if (os != NULL && os->type() == elfcpp::SHT_ARM_EXIDX) { Arm_output_section* exidx_output_section = Arm_output_section::as_arm_output_section(os); this->fix_exidx_coverage(layout, exidx_output_section, symtab); done_exidx_fixup = true; } } // The Cortex-A8 stubs are sensitive to layout of code sections. At the // beginning of each relaxation pass, just blow away all the stubs. // Alternatively, we could selectively remove only the stubs and reloc // information for code sections that have moved since the last pass. // That would require more book-keeping. typedef typename Stub_table_list::iterator Stub_table_iterator; if (this->fix_cortex_a8_) { // Clear all Cortex-A8 reloc information. for (typename Cortex_a8_relocs_info::const_iterator p = this->cortex_a8_relocs_info_.begin(); p != this->cortex_a8_relocs_info_.end(); ++p) delete p->second; this->cortex_a8_relocs_info_.clear(); // Remove all Cortex-A8 stubs. for (Stub_table_iterator sp = this->stub_tables_.begin(); sp != this->stub_tables_.end(); ++sp) (*sp)->remove_all_cortex_a8_stubs(); } // Scan relocs for relocation stubs for (Input_objects::Relobj_iterator op = input_objects->relobj_begin(); op != input_objects->relobj_end(); ++op) { Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(*op); arm_relobj->scan_sections_for_stubs(this, symtab, layout); } // Check all stub tables to see if any of them have their data sizes // or addresses alignments changed. These are the only things that // matter. bool any_stub_table_changed = false; Unordered_set sections_needing_adjustment; for (Stub_table_iterator sp = this->stub_tables_.begin(); (sp != this->stub_tables_.end()) && !any_stub_table_changed; ++sp) { if ((*sp)->update_data_size_and_addralign()) { // Update data size of stub table owner. Arm_input_section* owner = (*sp)->owner(); uint64_t address = owner->address(); off_t offset = owner->offset(); owner->reset_address_and_file_offset(); owner->set_address_and_file_offset(address, offset); sections_needing_adjustment.insert(owner->output_section()); any_stub_table_changed = true; } } // Output_section_data::output_section() returns a const pointer but we // need to update output sections, so we record all output sections needing // update above and scan the sections here to find out what sections need // to be updated. for(Layout::Section_list::const_iterator p = layout->section_list().begin(); p != layout->section_list().end(); ++p) { if (sections_needing_adjustment.find(*p) != sections_needing_adjustment.end()) (*p)->set_section_offsets_need_adjustment(); } // Stop relaxation if no EXIDX fix-up and no stub table change. bool continue_relaxation = done_exidx_fixup || any_stub_table_changed; // Finalize the stubs in the last relaxation pass. if (!continue_relaxation) { for (Stub_table_iterator sp = this->stub_tables_.begin(); (sp != this->stub_tables_.end()) && !any_stub_table_changed; ++sp) (*sp)->finalize_stubs(); // Update output local symbol counts of objects if necessary. for (Input_objects::Relobj_iterator op = input_objects->relobj_begin(); op != input_objects->relobj_end(); ++op) { Arm_relobj* arm_relobj = Arm_relobj::as_arm_relobj(*op); // Update output local symbol counts. We need to discard local // symbols defined in parts of input sections that are discarded by // relaxation. if (arm_relobj->output_local_symbol_count_needs_update()) arm_relobj->update_output_local_symbol_count(); } } return continue_relaxation; } // Relocate a stub. template void Target_arm::relocate_stub( Stub* stub, const Relocate_info<32, big_endian>* relinfo, Output_section* output_section, unsigned char* view, Arm_address address, section_size_type view_size) { Relocate relocate; const Stub_template* stub_template = stub->stub_template(); for (size_t i = 0; i < stub_template->reloc_count(); i++) { size_t reloc_insn_index = stub_template->reloc_insn_index(i); const Insn_template* insn = &stub_template->insns()[reloc_insn_index]; unsigned int r_type = insn->r_type(); section_size_type reloc_offset = stub_template->reloc_offset(i); section_size_type reloc_size = insn->size(); gold_assert(reloc_offset + reloc_size <= view_size); // This is the address of the stub destination. Arm_address target = stub->reloc_target(i) + insn->reloc_addend(); Symbol_value<32> symval; symval.set_output_value(target); // Synthesize a fake reloc just in case. We don't have a symbol so // we use 0. unsigned char reloc_buffer[elfcpp::Elf_sizes<32>::rel_size]; memset(reloc_buffer, 0, sizeof(reloc_buffer)); elfcpp::Rel_write<32, big_endian> reloc_write(reloc_buffer); reloc_write.put_r_offset(reloc_offset); reloc_write.put_r_info(elfcpp::elf_r_info<32>(0, r_type)); elfcpp::Rel<32, big_endian> rel(reloc_buffer); relocate.relocate(relinfo, this, output_section, this->fake_relnum_for_stubs, rel, r_type, NULL, &symval, view + reloc_offset, address + reloc_offset, reloc_size); } } // Determine whether an object attribute tag takes an integer, a // string or both. template int Target_arm::do_attribute_arg_type(int tag) const { if (tag == Object_attribute::Tag_compatibility) return (Object_attribute::ATTR_TYPE_FLAG_INT_VAL | Object_attribute::ATTR_TYPE_FLAG_STR_VAL); else if (tag == elfcpp::Tag_nodefaults) return (Object_attribute::ATTR_TYPE_FLAG_INT_VAL | Object_attribute::ATTR_TYPE_FLAG_NO_DEFAULT); else if (tag == elfcpp::Tag_CPU_raw_name || tag == elfcpp::Tag_CPU_name) return Object_attribute::ATTR_TYPE_FLAG_STR_VAL; else if (tag < 32) return Object_attribute::ATTR_TYPE_FLAG_INT_VAL; else return ((tag & 1) != 0 ? Object_attribute::ATTR_TYPE_FLAG_STR_VAL : Object_attribute::ATTR_TYPE_FLAG_INT_VAL); } // Reorder attributes. // // The ABI defines that Tag_conformance should be emitted first, and that // Tag_nodefaults should be second (if either is defined). This sets those // two positions, and bumps up the position of all the remaining tags to // compensate. template int Target_arm::do_attributes_order(int num) const { // Reorder the known object attributes in output. We want to move // Tag_conformance to position 4 and Tag_conformance to position 5 // and shift eveything between 4 .. Tag_conformance - 1 to make room. if (num == 4) return elfcpp::Tag_conformance; if (num == 5) return elfcpp::Tag_nodefaults; if ((num - 2) < elfcpp::Tag_nodefaults) return num - 2; if ((num - 1) < elfcpp::Tag_conformance) return num - 1; return num; } // Scan a span of THUMB code for Cortex-A8 erratum. template void Target_arm::scan_span_for_cortex_a8_erratum( Arm_relobj* arm_relobj, unsigned int shndx, section_size_type span_start, section_size_type span_end, const unsigned char* view, Arm_address address) { // Scan for 32-bit Thumb-2 branches which span two 4K regions, where: // // The opcode is BLX.W, BL.W, B.W, Bcc.W // The branch target is in the same 4KB region as the // first half of the branch. // The instruction before the branch is a 32-bit // length non-branch instruction. section_size_type i = span_start; bool last_was_32bit = false; bool last_was_branch = false; while (i < span_end) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; const Valtype* wv = reinterpret_cast(view + i); uint32_t insn = elfcpp::Swap<16, big_endian>::readval(wv); bool is_blx = false, is_b = false; bool is_bl = false, is_bcc = false; bool insn_32bit = (insn & 0xe000) == 0xe000 && (insn & 0x1800) != 0x0000; if (insn_32bit) { // Load the rest of the insn (in manual-friendly order). insn = (insn << 16) | elfcpp::Swap<16, big_endian>::readval(wv + 1); // Encoding T4: B.W. is_b = (insn & 0xf800d000U) == 0xf0009000U; // Encoding T1: BL.W. is_bl = (insn & 0xf800d000U) == 0xf000d000U; // Encoding T2: BLX.W. is_blx = (insn & 0xf800d000U) == 0xf000c000U; // Encoding T3: B.W (not permitted in IT block). is_bcc = ((insn & 0xf800d000U) == 0xf0008000U && (insn & 0x07f00000U) != 0x03800000U); } bool is_32bit_branch = is_b || is_bl || is_blx || is_bcc; // If this instruction is a 32-bit THUMB branch that crosses a 4K // page boundary and it follows 32-bit non-branch instruction, // we need to work around. if (is_32bit_branch && ((address + i) & 0xfffU) == 0xffeU && last_was_32bit && !last_was_branch) { // Check to see if there is a relocation stub for this branch. bool force_target_arm = false; bool force_target_thumb = false; const Cortex_a8_reloc* cortex_a8_reloc = NULL; Cortex_a8_relocs_info::const_iterator p = this->cortex_a8_relocs_info_.find(address + i); if (p != this->cortex_a8_relocs_info_.end()) { cortex_a8_reloc = p->second; bool target_is_thumb = (cortex_a8_reloc->destination() & 1) != 0; if (cortex_a8_reloc->r_type() == elfcpp::R_ARM_THM_CALL && !target_is_thumb) force_target_arm = true; else if (cortex_a8_reloc->r_type() == elfcpp::R_ARM_THM_CALL && target_is_thumb) force_target_thumb = true; } off_t offset; Stub_type stub_type = arm_stub_none; // Check if we have an offending branch instruction. uint16_t upper_insn = (insn >> 16) & 0xffffU; uint16_t lower_insn = insn & 0xffffU; typedef struct Arm_relocate_functions RelocFuncs; if (cortex_a8_reloc != NULL && cortex_a8_reloc->reloc_stub() != NULL) // We've already made a stub for this instruction, e.g. // it's a long branch or a Thumb->ARM stub. Assume that // stub will suffice to work around the A8 erratum (see // setting of always_after_branch above). ; else if (is_bcc) { offset = RelocFuncs::thumb32_cond_branch_offset(upper_insn, lower_insn); stub_type = arm_stub_a8_veneer_b_cond; } else if (is_b || is_bl || is_blx) { offset = RelocFuncs::thumb32_branch_offset(upper_insn, lower_insn); if (is_blx) offset &= ~3; stub_type = (is_blx ? arm_stub_a8_veneer_blx : (is_bl ? arm_stub_a8_veneer_bl : arm_stub_a8_veneer_b)); } if (stub_type != arm_stub_none) { Arm_address pc_for_insn = address + i + 4; // The original instruction is a BL, but the target is // an ARM instruction. If we were not making a stub, // the BL would have been converted to a BLX. Use the // BLX stub instead in that case. if (this->may_use_blx() && force_target_arm && stub_type == arm_stub_a8_veneer_bl) { stub_type = arm_stub_a8_veneer_blx; is_blx = true; is_bl = false; } // Conversely, if the original instruction was // BLX but the target is Thumb mode, use the BL stub. else if (force_target_thumb && stub_type == arm_stub_a8_veneer_blx) { stub_type = arm_stub_a8_veneer_bl; is_blx = false; is_bl = true; } if (is_blx) pc_for_insn &= ~3; // If we found a relocation, use the proper destination, // not the offset in the (unrelocated) instruction. // Note this is always done if we switched the stub type above. if (cortex_a8_reloc != NULL) offset = (off_t) (cortex_a8_reloc->destination() - pc_for_insn); Arm_address target = (pc_for_insn + offset) | (is_blx ? 0 : 1); // Add a new stub if destination address in in the same page. if (((address + i) & ~0xfffU) == (target & ~0xfffU)) { Cortex_a8_stub* stub = this->stub_factory_.make_cortex_a8_stub(stub_type, arm_relobj, shndx, address + i, target, insn); Stub_table* stub_table = arm_relobj->stub_table(shndx); gold_assert(stub_table != NULL); stub_table->add_cortex_a8_stub(address + i, stub); } } } i += insn_32bit ? 4 : 2; last_was_32bit = insn_32bit; last_was_branch = is_32bit_branch; } } // Apply the Cortex-A8 workaround. template void Target_arm::apply_cortex_a8_workaround( const Cortex_a8_stub* stub, Arm_address stub_address, unsigned char* insn_view, Arm_address insn_address) { typedef typename elfcpp::Swap<16, big_endian>::Valtype Valtype; Valtype* wv = reinterpret_cast(insn_view); Valtype upper_insn = elfcpp::Swap<16, big_endian>::readval(wv); Valtype lower_insn = elfcpp::Swap<16, big_endian>::readval(wv + 1); off_t branch_offset = stub_address - (insn_address + 4); typedef struct Arm_relocate_functions RelocFuncs; switch (stub->stub_template()->type()) { case arm_stub_a8_veneer_b_cond: gold_assert(!utils::has_overflow<21>(branch_offset)); upper_insn = RelocFuncs::thumb32_cond_branch_upper(upper_insn, branch_offset); lower_insn = RelocFuncs::thumb32_cond_branch_lower(lower_insn, branch_offset); break; case arm_stub_a8_veneer_b: case arm_stub_a8_veneer_bl: case arm_stub_a8_veneer_blx: if ((lower_insn & 0x5000U) == 0x4000U) // For a BLX instruction, make sure that the relocation is // rounded up to a word boundary. This follows the semantics of // the instruction which specifies that bit 1 of the target // address will come from bit 1 of the base address. branch_offset = (branch_offset + 2) & ~3; // Put BRANCH_OFFSET back into the insn. gold_assert(!utils::has_overflow<25>(branch_offset)); upper_insn = RelocFuncs::thumb32_branch_upper(upper_insn, branch_offset); lower_insn = RelocFuncs::thumb32_branch_lower(lower_insn, branch_offset); break; default: gold_unreachable(); } // Put the relocated value back in the object file: elfcpp::Swap<16, big_endian>::writeval(wv, upper_insn); elfcpp::Swap<16, big_endian>::writeval(wv + 1, lower_insn); } template class Target_selector_arm : public Target_selector { public: Target_selector_arm() : Target_selector(elfcpp::EM_ARM, 32, big_endian, (big_endian ? "elf32-bigarm" : "elf32-littlearm")) { } Target* do_instantiate_target() { return new Target_arm(); } }; // Fix .ARM.exidx section coverage. template void Target_arm::fix_exidx_coverage( Layout* layout, Arm_output_section* exidx_section, Symbol_table* symtab) { // We need to look at all the input sections in output in ascending // order of of output address. We do that by building a sorted list // of output sections by addresses. Then we looks at the output sections // in order. The input sections in an output section are already sorted // by addresses within the output section. typedef std::set Sorted_output_section_list; Sorted_output_section_list sorted_output_sections; Layout::Section_list section_list; layout->get_allocated_sections(§ion_list); for (Layout::Section_list::const_iterator p = section_list.begin(); p != section_list.end(); ++p) { // We only care about output sections that contain executable code. if (((*p)->flags() & elfcpp::SHF_EXECINSTR) != 0) sorted_output_sections.insert(*p); } // Go over the output sections in ascending order of output addresses. typedef typename Arm_output_section::Text_section_list Text_section_list; Text_section_list sorted_text_sections; for(typename Sorted_output_section_list::iterator p = sorted_output_sections.begin(); p != sorted_output_sections.end(); ++p) { Arm_output_section* arm_output_section = Arm_output_section::as_arm_output_section(*p); arm_output_section->append_text_sections_to_list(&sorted_text_sections); } exidx_section->fix_exidx_coverage(sorted_text_sections, symtab); } Target_selector_arm target_selector_arm; Target_selector_arm target_selector_armbe; } // End anonymous namespace.