| /* |
| * Copyright (C) 2019 The Android Open Source Project |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in |
| * the documentation and/or other materials provided with the |
| * distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS |
| * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE |
| * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, |
| * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, |
| * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS |
| * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED |
| * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
| * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT |
| * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF |
| * SUCH DAMAGE. |
| */ |
| |
| #include "linker_relocate.h" |
| |
| #include <elf.h> |
| #include <link.h> |
| |
| #include <type_traits> |
| |
| #include "linker.h" |
| #include "linker_debug.h" |
| #include "linker_globals.h" |
| #include "linker_gnu_hash.h" |
| #include "linker_phdr.h" |
| #include "linker_relocs.h" |
| #include "linker_reloc_iterators.h" |
| #include "linker_sleb128.h" |
| #include "linker_soinfo.h" |
| #include "private/bionic_globals.h" |
| |
| #include <platform/bionic/mte.h> |
| |
| static bool is_tls_reloc(ElfW(Word) type) { |
| switch (type) { |
| case R_GENERIC_TLS_DTPMOD: |
| case R_GENERIC_TLS_DTPREL: |
| case R_GENERIC_TLS_TPREL: |
| #if defined(R_GENERIC_TLSDESC) |
| case R_GENERIC_TLSDESC: |
| #endif |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| class Relocator { |
| public: |
| Relocator(const VersionTracker& version_tracker, const SymbolLookupList& lookup_list) |
| : version_tracker(version_tracker), lookup_list(lookup_list) |
| {} |
| |
| soinfo* si = nullptr; |
| const char* si_strtab = nullptr; |
| size_t si_strtab_size = 0; |
| ElfW(Sym)* si_symtab = nullptr; |
| |
| const VersionTracker& version_tracker; |
| const SymbolLookupList& lookup_list; |
| |
| // Cache key |
| ElfW(Word) cache_sym_val = 0; |
| // Cache value |
| const ElfW(Sym)* cache_sym = nullptr; |
| soinfo* cache_si = nullptr; |
| |
| std::vector<TlsDynamicResolverArg>* tlsdesc_args; |
| std::vector<std::pair<TlsDescriptor*, size_t>> deferred_tlsdesc_relocs; |
| size_t tls_tp_base = 0; |
| |
| __attribute__((always_inline)) |
| const char* get_string(ElfW(Word) index) { |
| if (__predict_false(index >= si_strtab_size)) { |
| async_safe_fatal("%s: strtab out of bounds error; STRSZ=%zd, name=%d", |
| si->get_realpath(), si_strtab_size, index); |
| } |
| return si_strtab + index; |
| } |
| }; |
| |
| template <bool DoLogging> |
| __attribute__((always_inline)) |
| static inline bool lookup_symbol(Relocator& relocator, uint32_t r_sym, const char* sym_name, |
| soinfo** found_in, const ElfW(Sym)** sym) { |
| if (r_sym == relocator.cache_sym_val) { |
| *found_in = relocator.cache_si; |
| *sym = relocator.cache_sym; |
| count_relocation_if<DoLogging>(kRelocSymbolCached); |
| } else { |
| const version_info* vi = nullptr; |
| if (!relocator.si->lookup_version_info(relocator.version_tracker, r_sym, sym_name, &vi)) { |
| return false; |
| } |
| |
| soinfo* local_found_in = nullptr; |
| const ElfW(Sym)* local_sym = soinfo_do_lookup(sym_name, vi, &local_found_in, relocator.lookup_list); |
| |
| relocator.cache_sym_val = r_sym; |
| relocator.cache_si = local_found_in; |
| relocator.cache_sym = local_sym; |
| *found_in = local_found_in; |
| *sym = local_sym; |
| } |
| |
| if (*sym == nullptr) { |
| if (ELF_ST_BIND(relocator.si_symtab[r_sym].st_info) != STB_WEAK) { |
| DL_ERR("cannot locate symbol \"%s\" referenced by \"%s\"...", sym_name, relocator.si->get_realpath()); |
| return false; |
| } |
| } |
| |
| count_relocation_if<DoLogging>(kRelocSymbol); |
| return true; |
| } |
| |
| enum class RelocMode { |
| // Fast path for JUMP_SLOT relocations. |
| JumpTable, |
| // Fast path for typical relocations: ABSOLUTE, GLOB_DAT, or RELATIVE. |
| Typical, |
| // Handle all relocation types, relocations in text sections, and statistics/tracing. |
| General, |
| }; |
| |
| struct linker_stats_t { |
| int count[kRelocMax]; |
| }; |
| |
| static linker_stats_t linker_stats; |
| |
| void count_relocation(RelocationKind kind) { |
| ++linker_stats.count[kind]; |
| } |
| |
| void print_linker_stats() { |
| LD_DEBUG(statistics, |
| "RELO STATS: %s: %d abs, %d rel, %d symbol (%d cached)", |
| g_argv[0], |
| linker_stats.count[kRelocAbsolute], |
| linker_stats.count[kRelocRelative], |
| linker_stats.count[kRelocSymbol], |
| linker_stats.count[kRelocSymbolCached]); |
| } |
| |
| static bool process_relocation_general(Relocator& relocator, const rel_t& reloc); |
| |
| template <RelocMode Mode> |
| __attribute__((always_inline)) |
| static bool process_relocation_impl(Relocator& relocator, const rel_t& reloc) { |
| constexpr bool IsGeneral = Mode == RelocMode::General; |
| |
| void* const rel_target = reinterpret_cast<void*>( |
| relocator.si->apply_memtag_if_mte_globals(reloc.r_offset + relocator.si->load_bias)); |
| const uint32_t r_type = ELFW(R_TYPE)(reloc.r_info); |
| const uint32_t r_sym = ELFW(R_SYM)(reloc.r_info); |
| |
| soinfo* found_in = nullptr; |
| const ElfW(Sym)* sym = nullptr; |
| const char* sym_name = nullptr; |
| ElfW(Addr) sym_addr = 0; |
| |
| if (r_sym != 0) { |
| sym_name = relocator.get_string(relocator.si_symtab[r_sym].st_name); |
| } |
| |
| // While relocating a DSO with text relocations (obsolete and 32-bit only), the .text segment is |
| // writable (but not executable). To call an ifunc, temporarily remap the segment as executable |
| // (but not writable). Then switch it back to continue applying relocations in the segment. |
| #if defined(__LP64__) |
| const bool handle_text_relocs = false; |
| auto protect_segments = []() { return true; }; |
| auto unprotect_segments = []() { return true; }; |
| #else |
| const bool handle_text_relocs = IsGeneral && relocator.si->has_text_relocations; |
| auto protect_segments = [&]() { |
| // Make .text executable. |
| if (phdr_table_protect_segments(relocator.si->phdr, relocator.si->phnum, |
| relocator.si->load_bias, relocator.si->should_pad_segments(), |
| relocator.si->should_use_16kib_app_compat()) < 0) { |
| DL_ERR("can't protect segments for \"%s\": %m", relocator.si->get_realpath()); |
| return false; |
| } |
| return true; |
| }; |
| auto unprotect_segments = [&]() { |
| // Make .text writable. |
| if (phdr_table_unprotect_segments(relocator.si->phdr, relocator.si->phnum, |
| relocator.si->load_bias, relocator.si->should_pad_segments(), |
| relocator.si->should_use_16kib_app_compat()) < 0) { |
| DL_ERR("can't unprotect loadable segments for \"%s\": %m", |
| relocator.si->get_realpath()); |
| return false; |
| } |
| return true; |
| }; |
| #endif |
| |
| // Skip symbol lookup for R_GENERIC_NONE relocations. |
| if (__predict_false(r_type == R_GENERIC_NONE)) { |
| LD_DEBUG(reloc && IsGeneral, "RELO NONE"); |
| return true; |
| } |
| |
| #if defined(USE_RELA) |
| auto get_addend_rel = [&]() -> ElfW(Addr) { return reloc.r_addend; }; |
| auto get_addend_norel = [&]() -> ElfW(Addr) { return reloc.r_addend; }; |
| #else |
| auto get_addend_rel = [&]() -> ElfW(Addr) { return *static_cast<ElfW(Addr)*>(rel_target); }; |
| auto get_addend_norel = [&]() -> ElfW(Addr) { return 0; }; |
| #endif |
| |
| if (!IsGeneral && __predict_false(is_tls_reloc(r_type))) { |
| // Always process TLS relocations using the slow code path, so that STB_LOCAL symbols are |
| // diagnosed, and ifunc processing is skipped. |
| return process_relocation_general(relocator, reloc); |
| } |
| |
| if (IsGeneral && is_tls_reloc(r_type)) { |
| if (r_sym == 0) { |
| // By convention in ld.bfd and lld, an omitted symbol on a TLS relocation |
| // is a reference to the current module. |
| found_in = relocator.si; |
| } else if (ELF_ST_BIND(relocator.si_symtab[r_sym].st_info) == STB_LOCAL) { |
| // In certain situations, the Gold linker accesses a TLS symbol using a |
| // relocation to an STB_LOCAL symbol in .dynsym of either STT_SECTION or |
| // STT_TLS type. Bionic doesn't support these relocations, so issue an |
| // error. References: |
| // - https://groups.google.com/d/topic/generic-abi/dJ4_Y78aQ2M/discussion |
| // - https://sourceware.org/bugzilla/show_bug.cgi?id=17699 |
| sym = &relocator.si_symtab[r_sym]; |
| auto sym_type = ELF_ST_TYPE(sym->st_info); |
| if (sym_type == STT_SECTION) { |
| DL_ERR("unexpected TLS reference to local section in \"%s\": sym type %d, rel type %u", |
| relocator.si->get_realpath(), sym_type, r_type); |
| } else { |
| DL_ERR( |
| "unexpected TLS reference to local symbol \"%s\" in \"%s\": sym type %d, rel type %u", |
| sym_name, relocator.si->get_realpath(), sym_type, r_type); |
| } |
| return false; |
| } else if (!lookup_symbol<IsGeneral>(relocator, r_sym, sym_name, &found_in, &sym)) { |
| return false; |
| } |
| if (found_in != nullptr && found_in->get_tls() == nullptr) { |
| // sym_name can be nullptr if r_sym is 0. A linker should never output an ELF file like this. |
| DL_ERR("TLS relocation refers to symbol \"%s\" in solib \"%s\" with no TLS segment", |
| sym_name, found_in->get_realpath()); |
| return false; |
| } |
| if (sym != nullptr) { |
| if (ELF_ST_TYPE(sym->st_info) != STT_TLS) { |
| // A toolchain should never output a relocation like this. |
| DL_ERR("reference to non-TLS symbol \"%s\" from TLS relocation in \"%s\"", |
| sym_name, relocator.si->get_realpath()); |
| return false; |
| } |
| sym_addr = sym->st_value; |
| } |
| } else { |
| if (r_sym == 0) { |
| // Do nothing. |
| } else { |
| if (!lookup_symbol<IsGeneral>(relocator, r_sym, sym_name, &found_in, &sym)) return false; |
| if (sym != nullptr) { |
| const bool should_protect_segments = handle_text_relocs && |
| found_in == relocator.si && |
| ELF_ST_TYPE(sym->st_info) == STT_GNU_IFUNC; |
| if (should_protect_segments && !protect_segments()) return false; |
| sym_addr = found_in->resolve_symbol_address(sym); |
| if (should_protect_segments && !unprotect_segments()) return false; |
| } else if constexpr (IsGeneral) { |
| // A weak reference to an undefined symbol. We typically use a zero symbol address, but |
| // use the relocation base for PC-relative relocations, so that the value written is zero. |
| switch (r_type) { |
| #if defined(__x86_64__) |
| case R_X86_64_PC32: |
| sym_addr = reinterpret_cast<ElfW(Addr)>(rel_target); |
| break; |
| #elif defined(__i386__) |
| case R_386_PC32: |
| sym_addr = reinterpret_cast<ElfW(Addr)>(rel_target); |
| break; |
| #endif |
| } |
| } |
| } |
| } |
| |
| if constexpr (IsGeneral || Mode == RelocMode::JumpTable) { |
| if (r_type == R_GENERIC_JUMP_SLOT) { |
| count_relocation_if<IsGeneral>(kRelocAbsolute); |
| const ElfW(Addr) result = sym_addr + get_addend_norel(); |
| LD_DEBUG(reloc && IsGeneral, "RELO JMP_SLOT %16p <- %16p %s", |
| rel_target, reinterpret_cast<void*>(result), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| return true; |
| } |
| } |
| |
| if constexpr (IsGeneral || Mode == RelocMode::Typical) { |
| // Almost all dynamic relocations are of one of these types, and most will be |
| // R_GENERIC_ABSOLUTE. The platform typically uses RELR instead, but R_GENERIC_RELATIVE is |
| // common in non-platform binaries. |
| if (r_type == R_GENERIC_ABSOLUTE) { |
| count_relocation_if<IsGeneral>(kRelocAbsolute); |
| if (found_in) sym_addr = found_in->apply_memtag_if_mte_globals(sym_addr); |
| const ElfW(Addr) result = sym_addr + get_addend_rel(); |
| LD_DEBUG(reloc && IsGeneral, "RELO ABSOLUTE %16p <- %16p %s", |
| rel_target, reinterpret_cast<void*>(result), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| return true; |
| } else if (r_type == R_GENERIC_GLOB_DAT) { |
| // The i386 psABI specifies that R_386_GLOB_DAT doesn't have an addend. The ARM ELF ABI |
| // document (IHI0044F) specifies that R_ARM_GLOB_DAT has an addend, but Bionic isn't adding |
| // it. |
| count_relocation_if<IsGeneral>(kRelocAbsolute); |
| if (found_in) sym_addr = found_in->apply_memtag_if_mte_globals(sym_addr); |
| const ElfW(Addr) result = sym_addr + get_addend_norel(); |
| LD_DEBUG(reloc && IsGeneral, "RELO GLOB_DAT %16p <- %16p %s", |
| rel_target, reinterpret_cast<void*>(result), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| return true; |
| } else if (r_type == R_GENERIC_RELATIVE) { |
| // In practice, r_sym is always zero, but if it weren't, the linker would still look up the |
| // referenced symbol (and abort if the symbol isn't found), even though it isn't used. |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| ElfW(Addr) result = relocator.si->load_bias + get_addend_rel(); |
| // MTE globals reuses the place bits for additional tag-derivation metadata for |
| // R_AARCH64_RELATIVE relocations, which makes it incompatible with |
| // `-Wl,--apply-dynamic-relocs`. This is enforced by lld, however there's nothing stopping |
| // Android binaries (particularly prebuilts) from building with this linker flag if they're |
| // not built with MTE globals. Thus, don't use the new relocation semantics if this DSO |
| // doesn't have MTE globals. |
| if (relocator.si->should_tag_memtag_globals()) { |
| int64_t* place = static_cast<int64_t*>(rel_target); |
| int64_t offset = *place; |
| result = relocator.si->apply_memtag_if_mte_globals(result + offset) - offset; |
| } |
| LD_DEBUG(reloc && IsGeneral, "RELO RELATIVE %16p <- %16p", |
| rel_target, reinterpret_cast<void*>(result)); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| return true; |
| } |
| } |
| |
| if constexpr (!IsGeneral) { |
| // Almost all relocations are handled above. Handle the remaining relocations below, in a |
| // separate function call. The symbol lookup will be repeated, but the result should be served |
| // from the 1-symbol lookup cache. |
| return process_relocation_general(relocator, reloc); |
| } |
| |
| switch (r_type) { |
| case R_GENERIC_IRELATIVE: |
| // In the linker, ifuncs are called as soon as possible so that string functions work. We must |
| // not call them again. (e.g. On arm32, resolving an ifunc changes the meaning of the addend |
| // from a resolver function to the implementation.) |
| if (!relocator.si->is_linker()) { |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| const ElfW(Addr) ifunc_addr = relocator.si->load_bias + get_addend_rel(); |
| LD_DEBUG(reloc && IsGeneral, "RELO IRELATIVE %16p <- %16p", |
| rel_target, reinterpret_cast<void*>(ifunc_addr)); |
| if (handle_text_relocs && !protect_segments()) return false; |
| const ElfW(Addr) result = call_ifunc_resolver(ifunc_addr); |
| if (handle_text_relocs && !unprotect_segments()) return false; |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| } |
| break; |
| case R_GENERIC_COPY: |
| // Copy relocations allow read-only data or code in a non-PIE executable to access a |
| // variable from a DSO. The executable reserves extra space in its .bss section, and the |
| // linker copies the variable into the extra space. The executable then exports its copy |
| // to interpose the copy in the DSO. |
| // |
| // Bionic only supports PIE executables, so copy relocations aren't supported. The ARM and |
| // AArch64 ABI documents only allow them for ET_EXEC (non-PIE) objects. See IHI0056B and |
| // IHI0044F. |
| DL_ERR("%s COPY relocations are not supported", relocator.si->get_realpath()); |
| return false; |
| case R_GENERIC_TLS_TPREL: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| ElfW(Addr) tpoff = 0; |
| if (found_in == nullptr) { |
| // Unresolved weak relocation. Leave tpoff at 0 to resolve |
| // &weak_tls_symbol to __get_tls(). |
| } else { |
| CHECK(found_in->get_tls() != nullptr); // We rejected a missing TLS segment above. |
| const TlsModule& mod = get_tls_module(found_in->get_tls()->module_id); |
| if (mod.static_offset != SIZE_MAX) { |
| tpoff += mod.static_offset - relocator.tls_tp_base; |
| } else { |
| DL_ERR("TLS symbol \"%s\" in dlopened \"%s\" referenced from \"%s\" using IE access model", |
| sym_name, found_in->get_realpath(), relocator.si->get_realpath()); |
| return false; |
| } |
| } |
| tpoff += sym_addr + get_addend_rel(); |
| LD_DEBUG(reloc && IsGeneral, "RELO TLS_TPREL %16p <- %16p %s", |
| rel_target, reinterpret_cast<void*>(tpoff), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = tpoff; |
| } |
| break; |
| case R_GENERIC_TLS_DTPMOD: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| size_t module_id = 0; |
| if (found_in == nullptr) { |
| // Unresolved weak relocation. Evaluate the module ID to 0. |
| } else { |
| CHECK(found_in->get_tls() != nullptr); // We rejected a missing TLS segment above. |
| module_id = found_in->get_tls()->module_id; |
| CHECK(module_id != kTlsUninitializedModuleId); |
| } |
| LD_DEBUG(reloc && IsGeneral, "RELO TLS_DTPMOD %16p <- %zu %s", |
| rel_target, module_id, sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = module_id; |
| } |
| break; |
| case R_GENERIC_TLS_DTPREL: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| const ElfW(Addr) result = sym_addr + get_addend_rel() - TLS_DTV_OFFSET; |
| LD_DEBUG(reloc && IsGeneral, "RELO TLS_DTPREL %16p <- %16p %s", |
| rel_target, reinterpret_cast<void*>(result), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| } |
| break; |
| |
| #if defined(__aarch64__) || defined(__riscv) |
| // Bionic currently implements TLSDESC for arm64 and riscv64. This implementation should work |
| // with other architectures, as long as the resolver functions are implemented. |
| case R_GENERIC_TLSDESC: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| ElfW(Addr) addend = reloc.r_addend; |
| TlsDescriptor* desc = static_cast<TlsDescriptor*>(rel_target); |
| if (found_in == nullptr) { |
| // Unresolved weak relocation. |
| desc->func = tlsdesc_resolver_unresolved_weak; |
| desc->arg = addend; |
| LD_DEBUG(reloc && IsGeneral, "RELO TLSDESC %16p <- unresolved weak, addend 0x%zx %s", |
| rel_target, static_cast<size_t>(addend), sym_name); |
| } else { |
| CHECK(found_in->get_tls() != nullptr); // We rejected a missing TLS segment above. |
| size_t module_id = found_in->get_tls()->module_id; |
| const TlsModule& mod = get_tls_module(module_id); |
| if (mod.static_offset != SIZE_MAX) { |
| desc->func = tlsdesc_resolver_static; |
| desc->arg = mod.static_offset - relocator.tls_tp_base + sym_addr + addend; |
| LD_DEBUG(reloc && IsGeneral, "RELO TLSDESC %16p <- static (0x%zx - 0x%zx + 0x%zx + 0x%zx) %s", |
| rel_target, mod.static_offset, relocator.tls_tp_base, |
| static_cast<size_t>(sym_addr), static_cast<size_t>(addend), |
| sym_name); |
| } else { |
| relocator.tlsdesc_args->push_back({ |
| .generation = mod.first_generation, |
| .index.module_id = module_id, |
| .index.offset = sym_addr + addend, |
| }); |
| // Defer the TLSDESC relocation until the address of the TlsDynamicResolverArg object |
| // is finalized. |
| relocator.deferred_tlsdesc_relocs.push_back({ |
| desc, relocator.tlsdesc_args->size() - 1 |
| }); |
| const TlsDynamicResolverArg& desc_arg = relocator.tlsdesc_args->back(); |
| LD_DEBUG(reloc && IsGeneral, "RELO TLSDESC %16p <- dynamic (gen %zu, mod %zu, off %zu) %s", |
| rel_target, desc_arg.generation, desc_arg.index.module_id, |
| desc_arg.index.offset, sym_name); |
| } |
| } |
| } |
| break; |
| #endif // defined(__aarch64__) || defined(__riscv) |
| |
| #if defined(__x86_64__) |
| case R_X86_64_32: |
| count_relocation_if<IsGeneral>(kRelocAbsolute); |
| { |
| const Elf32_Addr result = sym_addr + reloc.r_addend; |
| LD_DEBUG(reloc && IsGeneral, "RELO R_X86_64_32 %16p <- 0x%08x %s", |
| rel_target, result, sym_name); |
| *static_cast<Elf32_Addr*>(rel_target) = result; |
| } |
| break; |
| case R_X86_64_PC32: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| const ElfW(Addr) target = sym_addr + reloc.r_addend; |
| const ElfW(Addr) base = reinterpret_cast<ElfW(Addr)>(rel_target); |
| const Elf32_Addr result = target - base; |
| LD_DEBUG(reloc && IsGeneral, "RELO R_X86_64_PC32 %16p <- 0x%08x (%16p - %16p) %s", |
| rel_target, result, reinterpret_cast<void*>(target), |
| reinterpret_cast<void*>(base), sym_name); |
| *static_cast<Elf32_Addr*>(rel_target) = result; |
| } |
| break; |
| #elif defined(__i386__) |
| case R_386_PC32: |
| count_relocation_if<IsGeneral>(kRelocRelative); |
| { |
| const ElfW(Addr) target = sym_addr + get_addend_rel(); |
| const ElfW(Addr) base = reinterpret_cast<ElfW(Addr)>(rel_target); |
| const ElfW(Addr) result = target - base; |
| LD_DEBUG(reloc && IsGeneral, "RELO R_386_PC32 %16p <- 0x%08x (%16p - %16p) %s", |
| rel_target, result, reinterpret_cast<void*>(target), |
| reinterpret_cast<void*>(base), sym_name); |
| *static_cast<ElfW(Addr)*>(rel_target) = result; |
| } |
| break; |
| #endif |
| default: |
| DL_ERR("unknown reloc type %d in \"%s\"", r_type, relocator.si->get_realpath()); |
| return false; |
| } |
| return true; |
| } |
| |
| __attribute__((noinline)) |
| static bool process_relocation_general(Relocator& relocator, const rel_t& reloc) { |
| return process_relocation_impl<RelocMode::General>(relocator, reloc); |
| } |
| |
| template <RelocMode Mode> |
| __attribute__((always_inline)) |
| static inline bool process_relocation(Relocator& relocator, const rel_t& reloc) { |
| return Mode == RelocMode::General ? |
| process_relocation_general(relocator, reloc) : |
| process_relocation_impl<Mode>(relocator, reloc); |
| } |
| |
| template <RelocMode Mode> |
| __attribute__((noinline)) |
| static bool plain_relocate_impl(Relocator& relocator, rel_t* rels, size_t rel_count) { |
| for (size_t i = 0; i < rel_count; ++i) { |
| if (!process_relocation<Mode>(relocator, rels[i])) { |
| return false; |
| } |
| } |
| return true; |
| } |
| |
| template <RelocMode Mode> |
| __attribute__((noinline)) |
| static bool packed_relocate_impl(Relocator& relocator, sleb128_decoder decoder) { |
| return for_all_packed_relocs(decoder, [&](const rel_t& reloc) { |
| return process_relocation<Mode>(relocator, reloc); |
| }); |
| } |
| |
| static bool needs_slow_relocate_loop(const Relocator& relocator __unused) { |
| #if !defined(__LP64__) |
| if (relocator.si->has_text_relocations) return true; |
| #endif |
| // Both LD_DEBUG relocation logging and statistics need the slow path. |
| if (g_linker_debug_config.any || g_linker_debug_config.statistics) { |
| return true; |
| } |
| return false; |
| } |
| |
| template <RelocMode OptMode, typename ...Args> |
| static bool plain_relocate(Relocator& relocator, Args ...args) { |
| return needs_slow_relocate_loop(relocator) ? |
| plain_relocate_impl<RelocMode::General>(relocator, args...) : |
| plain_relocate_impl<OptMode>(relocator, args...); |
| } |
| |
| template <RelocMode OptMode, typename ...Args> |
| static bool packed_relocate(Relocator& relocator, Args ...args) { |
| return needs_slow_relocate_loop(relocator) ? |
| packed_relocate_impl<RelocMode::General>(relocator, args...) : |
| packed_relocate_impl<OptMode>(relocator, args...); |
| } |
| |
| bool soinfo::relocate(const SymbolLookupList& lookup_list) { |
| // For ldd, don't apply relocations because TLS segments are not registered. |
| // We don't care whether ldd diagnoses unresolved symbols. |
| if (g_is_ldd) { |
| return true; |
| } |
| |
| VersionTracker version_tracker; |
| |
| if (!version_tracker.init(this)) { |
| return false; |
| } |
| |
| Relocator relocator(version_tracker, lookup_list); |
| relocator.si = this; |
| relocator.si_strtab = strtab_; |
| relocator.si_strtab_size = has_min_version(1) ? strtab_size_ : SIZE_MAX; |
| relocator.si_symtab = symtab_; |
| relocator.tlsdesc_args = &tlsdesc_args_; |
| relocator.tls_tp_base = __libc_shared_globals()->static_tls_layout.offset_thread_pointer(); |
| |
| // The linker already applied its RELR relocations in an earlier pass, so |
| // skip the RELR relocations for the linker. |
| if (relr_ != nullptr && !is_linker()) { |
| LD_DEBUG(reloc, "[ relocating %s relr ]", get_realpath()); |
| const ElfW(Relr)* begin = relr_; |
| const ElfW(Relr)* end = relr_ + relr_count_; |
| if (!relocate_relr(begin, end, load_bias, should_tag_memtag_globals())) { |
| return false; |
| } |
| } |
| |
| if (android_relocs_ != nullptr) { |
| // check signature |
| if (android_relocs_size_ > 3 && |
| android_relocs_[0] == 'A' && |
| android_relocs_[1] == 'P' && |
| android_relocs_[2] == 'S' && |
| android_relocs_[3] == '2') { |
| LD_DEBUG(reloc, "[ relocating %s android rel/rela ]", get_realpath()); |
| |
| const uint8_t* packed_relocs = android_relocs_ + 4; |
| const size_t packed_relocs_size = android_relocs_size_ - 4; |
| |
| if (!packed_relocate<RelocMode::Typical>(relocator, sleb128_decoder(packed_relocs, packed_relocs_size))) { |
| return false; |
| } |
| } else { |
| DL_ERR("bad android relocation header."); |
| return false; |
| } |
| } |
| |
| #if defined(USE_RELA) |
| if (rela_ != nullptr) { |
| LD_DEBUG(reloc, "[ relocating %s rela ]", get_realpath()); |
| |
| if (!plain_relocate<RelocMode::Typical>(relocator, rela_, rela_count_)) { |
| return false; |
| } |
| } |
| if (plt_rela_ != nullptr) { |
| LD_DEBUG(reloc, "[ relocating %s plt rela ]", get_realpath()); |
| if (!plain_relocate<RelocMode::JumpTable>(relocator, plt_rela_, plt_rela_count_)) { |
| return false; |
| } |
| } |
| #else |
| if (rel_ != nullptr) { |
| LD_DEBUG(reloc, "[ relocating %s rel ]", get_realpath()); |
| if (!plain_relocate<RelocMode::Typical>(relocator, rel_, rel_count_)) { |
| return false; |
| } |
| } |
| if (plt_rel_ != nullptr) { |
| LD_DEBUG(reloc, "[ relocating %s plt rel ]", get_realpath()); |
| if (!plain_relocate<RelocMode::JumpTable>(relocator, plt_rel_, plt_rel_count_)) { |
| return false; |
| } |
| } |
| #endif |
| |
| // Once the tlsdesc_args_ vector's size is finalized, we can write the addresses of its elements |
| // into the TLSDESC relocations. |
| #if defined(__aarch64__) || defined(__riscv) |
| // Bionic currently only implements TLSDESC for arm64 and riscv64. |
| for (const std::pair<TlsDescriptor*, size_t>& pair : relocator.deferred_tlsdesc_relocs) { |
| TlsDescriptor* desc = pair.first; |
| desc->func = tlsdesc_resolver_dynamic; |
| desc->arg = reinterpret_cast<size_t>(&tlsdesc_args_[pair.second]); |
| } |
| #endif // defined(__aarch64__) || defined(__riscv) |
| |
| return true; |
| } |