blob: 461a3312769b7e8aa81ccdfc6fbb4573687f43f2 [file] [log] [blame]
/*
* Copyright (C) 2015 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "linker/arm/relative_patcher_thumb2.h"
#include "arch/arm/instruction_set_features_arm.h"
#include "base/casts.h"
#include "driver/compiler_options.h"
#include "linker/relative_patcher_test.h"
#include "lock_word.h"
#include "mirror/array-inl.h"
#include "mirror/object.h"
#include "oat/oat_quick_method_header.h"
#include "optimizing/code_generator_arm_vixl.h"
#include "optimizing/optimizing_unit_test.h"
namespace art {
namespace linker {
class Thumb2RelativePatcherTest : public RelativePatcherTest {
public:
Thumb2RelativePatcherTest() : RelativePatcherTest(InstructionSet::kThumb2, "default") { }
protected:
static const uint8_t kCallRawCode[];
static const ArrayRef<const uint8_t> kCallCode;
static const uint8_t kNopRawCode[];
static const ArrayRef<const uint8_t> kNopCode;
static const uint8_t kUnpatchedPcRelativeRawCode[];
static const ArrayRef<const uint8_t> kUnpatchedPcRelativeCode;
static const uint32_t kPcInsnOffset;
// The PC in Thumb mode is 4 bytes after the instruction location.
static constexpr uint32_t kPcAdjustment = 4u;
// Branches within range [-256, 256) can be created from these by adding the low 8 bits.
static constexpr uint32_t kBlPlus0 = 0xf000f800u;
static constexpr uint32_t kBlMinus256 = 0xf7ffff00u;
// Special BL values.
static constexpr uint32_t kBlPlusMax = 0xf3ffd7ffu;
static constexpr uint32_t kBlMinusMax = 0xf400d000u;
// BNE +0, 32-bit, encoding T3. Bits 0-10, 11, 13, 16-21, 26 are placeholder for target offset.
static constexpr uint32_t kBneWPlus0 = 0xf0408000u;
// LDR immediate, 16-bit, encoding T1. Bits 6-10 are imm5, 0-2 are Rt, 3-5 are Rn.
static constexpr uint32_t kLdrInsn = 0x6800u;
// LDR immediate, 32-bit, encoding T3. Bits 0-11 are offset, 12-15 are Rt, 16-20 are Rn.
static constexpr uint32_t kLdrWInsn = 0xf8d00000u;
// LDR immediate, negative offset, encoding T4. Bits 0-7 are the offset to subtract.
static constexpr uint32_t kLdrNegativeOffset = 0xf8500c00u;
// LDR register, lsl #2. Bits 4-5 are the imm2, i.e. the lsl shift.
static constexpr uint32_t kLdrRegLsl2 = 0xf8500020u;
// NOP instructions.
static constexpr uint32_t kNopInsn = 0xbf00u;
static constexpr uint32_t kNopWInsn = 0xf3af8000u;
void InsertInsn(std::vector<uint8_t>* code, size_t pos, uint32_t insn) {
CHECK_LE(pos, code->size());
if (IsUint<16>(insn)) {
const uint8_t insn_code[] = {
static_cast<uint8_t>(insn),
static_cast<uint8_t>(insn >> 8),
};
static_assert(sizeof(insn_code) == 2u, "Invalid sizeof(insn_code).");
code->insert(code->begin() + pos, insn_code, insn_code + sizeof(insn_code));
} else {
const uint8_t insn_code[] = {
static_cast<uint8_t>(insn >> 16),
static_cast<uint8_t>(insn >> 24),
static_cast<uint8_t>(insn),
static_cast<uint8_t>(insn >> 8),
};
static_assert(sizeof(insn_code) == 4u, "Invalid sizeof(insn_code).");
code->insert(code->begin() + pos, insn_code, insn_code + sizeof(insn_code));
}
}
void PushBackInsn(std::vector<uint8_t>* code, uint32_t insn) {
InsertInsn(code, code->size(), insn);
}
std::vector<uint8_t> GenNops(size_t num_nops) {
std::vector<uint8_t> result;
result.reserve(num_nops * 2u);
for (size_t i = 0; i != num_nops; ++i) {
PushBackInsn(&result, kNopInsn);
}
return result;
}
std::vector<uint8_t> RawCode(std::initializer_list<uint32_t> insns) {
std::vector<uint8_t> raw_code;
size_t number_of_16_bit_insns =
std::count_if(insns.begin(), insns.end(), [](uint32_t x) { return IsUint<16>(x); });
raw_code.reserve(insns.size() * 4u - number_of_16_bit_insns * 2u);
for (uint32_t insn : insns) {
PushBackInsn(&raw_code, insn);
}
return raw_code;
}
uint32_t BneWWithOffset(uint32_t bne_offset, uint32_t target_offset) {
if (!IsAligned<2u>(bne_offset)) {
LOG(ERROR) << "Unaligned bne_offset: " << bne_offset;
return 0xffffffffu; // Fails code diff later.
}
if (!IsAligned<2u>(target_offset)) {
LOG(ERROR) << "Unaligned target_offset: " << target_offset;
return 0xffffffffu; // Fails code diff later.
}
uint32_t diff = target_offset - bne_offset - kPcAdjustment;
DCHECK_ALIGNED(diff, 2u);
if ((diff >> 20) != 0 && (diff >> 20) != 0xfffu) {
LOG(ERROR) << "Target out of range: " << diff;
return 0xffffffffu; // Fails code diff later.
}
return kBneWPlus0 | ((diff >> 1) & 0x7ffu) // imm11
| (((diff >> 12) & 0x3fu) << 16) // imm6
| (((diff >> 18) & 1) << 13) // J1
| (((diff >> 19) & 1) << 11) // J2
| (((diff >> 20) & 1) << 26); // S
}
uint32_t Create2MethodsWithGap(const ArrayRef<const uint8_t>& method1_code,
const ArrayRef<const LinkerPatch>& method1_patches,
const ArrayRef<const uint8_t>& last_method_code,
const ArrayRef<const LinkerPatch>& last_method_patches,
uint32_t distance_without_thunks) {
CHECK_EQ(distance_without_thunks % kArmCodeAlignment, 0u);
uint32_t method1_offset =
kTrampolineSize + CodeAlignmentSize(kTrampolineSize) + sizeof(OatQuickMethodHeader);
AddCompiledMethod(MethodRef(1u), method1_code, method1_patches);
const uint32_t gap_start = method1_offset + method1_code.size();
// We want to put the last method at a very precise offset.
const uint32_t last_method_offset = method1_offset + distance_without_thunks;
CHECK_ALIGNED(last_method_offset, kArmCodeAlignment);
const uint32_t gap_end = last_method_offset - sizeof(OatQuickMethodHeader);
// Fill the gap with intermediate methods in chunks of 2MiB and the first in [2MiB, 4MiB).
// (This allows deduplicating the small chunks to avoid using 32MiB of memory for +-16MiB
// offsets by this test. Making the first chunk bigger makes it easy to give all intermediate
// methods the same alignment of the end, so the thunk insertion adds a predictable size as
// long as it's after the first chunk.)
uint32_t method_idx = 2u;
constexpr uint32_t kSmallChunkSize = 2 * MB;
std::vector<uint8_t> gap_code;
uint32_t gap_size = gap_end - gap_start;
uint32_t num_small_chunks = std::max(gap_size / kSmallChunkSize, 1u) - 1u;
uint32_t chunk_start = gap_start;
uint32_t chunk_size = gap_size - num_small_chunks * kSmallChunkSize;
for (uint32_t i = 0; i <= num_small_chunks; ++i) { // num_small_chunks+1 iterations.
uint32_t chunk_code_size =
chunk_size - CodeAlignmentSize(chunk_start) - sizeof(OatQuickMethodHeader);
gap_code.resize(chunk_code_size, 0u);
AddCompiledMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(gap_code));
method_idx += 1u;
chunk_start += chunk_size;
chunk_size = kSmallChunkSize; // For all but the first chunk.
DCHECK_EQ(CodeAlignmentSize(gap_end), CodeAlignmentSize(chunk_start));
}
// Add the last method and link
AddCompiledMethod(MethodRef(method_idx), last_method_code, last_method_patches);
Link();
// Check assumptions.
CHECK_EQ(GetMethodOffset(1), method1_offset);
auto last_result = method_offset_map_.FindMethodOffset(MethodRef(method_idx));
CHECK(last_result.first);
// There may be a thunk before method2.
if (last_result.second != last_method_offset + 1 /* thumb mode */) {
// Thunk present. Check that there's only one.
uint32_t thunk_end =
CompiledCode::AlignCode(gap_end, InstructionSet::kThumb2) + MethodCallThunkSize();
uint32_t header_offset = thunk_end + CodeAlignmentSize(thunk_end);
CHECK_EQ(last_result.second,
header_offset + sizeof(OatQuickMethodHeader) + 1 /* thumb mode */);
}
return method_idx;
}
uint32_t GetMethodOffset(uint32_t method_idx) {
auto result = method_offset_map_.FindMethodOffset(MethodRef(method_idx));
CHECK(result.first);
CHECK_NE(result.second & 1u, 0u);
return result.second - 1 /* thumb mode */;
}
std::vector<uint8_t> CompileThunk(const LinkerPatch& patch,
/*out*/ std::string* debug_name = nullptr) {
OptimizingUnitTestHelper helper;
HGraph* graph = helper.CreateGraph();
CompilerOptions compiler_options;
// Set isa to Thumb2.
compiler_options.instruction_set_ = instruction_set_;
compiler_options.instruction_set_features_ =
InstructionSetFeatures::FromBitmap(instruction_set_, instruction_set_features_->AsBitmap());
CHECK(compiler_options.instruction_set_features_->Equals(instruction_set_features_.get()));
// If a test requests that implicit null checks are enabled or disabled,
// apply that option, otherwise use the default from `CompilerOptions`.
if (implicit_null_checks_.has_value()) {
compiler_options.implicit_null_checks_ = implicit_null_checks_.value();
}
arm::CodeGeneratorARMVIXL codegen(graph, compiler_options);
ArenaVector<uint8_t> code(helper.GetAllocator()->Adapter());
codegen.EmitThunkCode(patch, &code, debug_name);
return std::vector<uint8_t>(code.begin(), code.end());
}
void AddCompiledMethod(
MethodReference method_ref,
const ArrayRef<const uint8_t>& code,
const ArrayRef<const LinkerPatch>& patches = ArrayRef<const LinkerPatch>()) {
RelativePatcherTest::AddCompiledMethod(method_ref, code, patches);
// Make sure the ThunkProvider has all the necessary thunks.
for (const LinkerPatch& patch : patches) {
if (patch.GetType() == LinkerPatch::Type::kCallEntrypoint ||
patch.GetType() == LinkerPatch::Type::kBakerReadBarrierBranch ||
patch.GetType() == LinkerPatch::Type::kCallRelative) {
std::string debug_name;
std::vector<uint8_t> thunk_code = CompileThunk(patch, &debug_name);
thunk_provider_.SetThunkCode(patch, ArrayRef<const uint8_t>(thunk_code), debug_name);
}
}
}
std::vector<uint8_t> CompileMethodCallThunk() {
LinkerPatch patch = LinkerPatch::RelativeCodePatch(/* literal_offset */ 0u,
/* target_dex_file*/ nullptr,
/* target_method_idx */ 0u);
return CompileThunk(patch);
}
uint32_t MethodCallThunkSize() {
return CompileMethodCallThunk().size();
}
bool CheckThunk(uint32_t thunk_offset) {
const std::vector<uint8_t> expected_code = CompileMethodCallThunk();
if (output_.size() < thunk_offset + expected_code.size()) {
LOG(ERROR) << "output_.size() == " << output_.size() << " < "
<< "thunk_offset + expected_code.size() == " << (thunk_offset + expected_code.size());
return false;
}
ArrayRef<const uint8_t> linked_code(&output_[thunk_offset], expected_code.size());
if (linked_code == ArrayRef<const uint8_t>(expected_code)) {
return true;
}
// Log failure info.
DumpDiff(ArrayRef<const uint8_t>(expected_code), linked_code);
return false;
}
std::vector<uint8_t> GenNopsAndBl(size_t num_nops, uint32_t bl) {
std::vector<uint8_t> result;
result.reserve(num_nops * 2u + 4u);
for (size_t i = 0; i != num_nops; ++i) {
PushBackInsn(&result, kNopInsn);
}
PushBackInsn(&result, bl);
return result;
}
void TestStringBssEntry(uint32_t bss_begin, uint32_t string_entry_offset);
void TestStringReference(uint32_t string_offset);
void CheckPcRelativePatch(const ArrayRef<const LinkerPatch>& patches, uint32_t target_offset);
static uint32_t EncodeBakerReadBarrierFieldData(uint32_t base_reg,
uint32_t holder_reg,
bool narrow) {
return arm::CodeGeneratorARMVIXL::EncodeBakerReadBarrierFieldData(base_reg, holder_reg, narrow);
}
static uint32_t EncodeBakerReadBarrierArrayData(uint32_t base_reg) {
return arm::CodeGeneratorARMVIXL::EncodeBakerReadBarrierArrayData(base_reg);
}
static uint32_t EncodeBakerReadBarrierGcRootData(uint32_t root_reg, bool narrow) {
return arm::CodeGeneratorARMVIXL::EncodeBakerReadBarrierGcRootData(root_reg, narrow);
}
std::vector<uint8_t> CompileBakerOffsetThunk(uint32_t base_reg,
uint32_t holder_reg,
bool narrow) {
const LinkerPatch patch = LinkerPatch::BakerReadBarrierBranchPatch(
/* literal_offset */ 0u, EncodeBakerReadBarrierFieldData(base_reg, holder_reg, narrow));
return CompileThunk(patch);
}
std::vector<uint8_t> CompileBakerArrayThunk(uint32_t base_reg) {
LinkerPatch patch = LinkerPatch::BakerReadBarrierBranchPatch(
/* literal_offset */ 0u, EncodeBakerReadBarrierArrayData(base_reg));
return CompileThunk(patch);
}
std::vector<uint8_t> CompileBakerGcRootThunk(uint32_t root_reg, bool narrow) {
LinkerPatch patch = LinkerPatch::BakerReadBarrierBranchPatch(
/* literal_offset */ 0u, EncodeBakerReadBarrierGcRootData(root_reg, narrow));
return CompileThunk(patch);
}
uint32_t GetOutputInsn32(uint32_t offset) {
CHECK_LE(offset, output_.size());
CHECK_GE(output_.size() - offset, 4u);
return (static_cast<uint32_t>(output_[offset]) << 16) |
(static_cast<uint32_t>(output_[offset + 1]) << 24) |
(static_cast<uint32_t>(output_[offset + 2]) << 0) |
(static_cast<uint32_t>(output_[offset + 3]) << 8);
}
uint16_t GetOutputInsn16(uint32_t offset) {
CHECK_LE(offset, output_.size());
CHECK_GE(output_.size() - offset, 2u);
return (static_cast<uint32_t>(output_[offset]) << 0) |
(static_cast<uint32_t>(output_[offset + 1]) << 8);
}
void TestBakerFieldWide(uint32_t offset, uint32_t ref_reg, bool implicit_null_checks);
void TestBakerFieldNarrow(uint32_t offset, uint32_t ref_reg, bool implicit_null_checks);
void Reset() final {
RelativePatcherTest::Reset();
implicit_null_checks_ = std::nullopt;
}
private:
std::optional<bool> implicit_null_checks_ = std::nullopt;
};
const uint8_t Thumb2RelativePatcherTest::kCallRawCode[] = {
0x00, 0xf0, 0x00, 0xf8
};
const ArrayRef<const uint8_t> Thumb2RelativePatcherTest::kCallCode(kCallRawCode);
const uint8_t Thumb2RelativePatcherTest::kNopRawCode[] = {
0x00, 0xbf
};
const ArrayRef<const uint8_t> Thumb2RelativePatcherTest::kNopCode(kNopRawCode);
const uint8_t Thumb2RelativePatcherTest::kUnpatchedPcRelativeRawCode[] = {
0x40, 0xf2, 0x00, 0x00, // MOVW r0, #0 (placeholder)
0xc0, 0xf2, 0x00, 0x00, // MOVT r0, #0 (placeholder)
0x78, 0x44, // ADD r0, pc
};
const ArrayRef<const uint8_t> Thumb2RelativePatcherTest::kUnpatchedPcRelativeCode(
kUnpatchedPcRelativeRawCode);
const uint32_t Thumb2RelativePatcherTest::kPcInsnOffset = 8u;
void Thumb2RelativePatcherTest::TestStringBssEntry(uint32_t bss_begin,
uint32_t string_entry_offset) {
constexpr uint32_t kStringIndex = 1u;
string_index_to_offset_map_.Put(kStringIndex, string_entry_offset);
bss_begin_ = bss_begin;
const LinkerPatch patches[] = {
LinkerPatch::StringBssEntryPatch(0u, nullptr, kPcInsnOffset, kStringIndex),
LinkerPatch::StringBssEntryPatch(4u, nullptr, kPcInsnOffset, kStringIndex),
};
CheckPcRelativePatch(ArrayRef<const LinkerPatch>(patches), bss_begin_ + string_entry_offset);
}
void Thumb2RelativePatcherTest::TestStringReference(uint32_t string_offset) {
constexpr uint32_t kStringIndex = 1u;
string_index_to_offset_map_.Put(kStringIndex, string_offset);
const LinkerPatch patches[] = {
LinkerPatch::RelativeStringPatch(0u, nullptr, kPcInsnOffset, kStringIndex),
LinkerPatch::RelativeStringPatch(4u, nullptr, kPcInsnOffset, kStringIndex),
};
CheckPcRelativePatch(ArrayRef<const LinkerPatch>(patches), string_offset);
}
void Thumb2RelativePatcherTest::CheckPcRelativePatch(const ArrayRef<const LinkerPatch>& patches,
uint32_t target_offset) {
AddCompiledMethod(MethodRef(1u), kUnpatchedPcRelativeCode, ArrayRef<const LinkerPatch>(patches));
Link();
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t pc_base_offset = method1_offset + kPcInsnOffset + 4u /* PC adjustment */;
uint32_t diff = target_offset - pc_base_offset;
// Distribute the bits of the diff between the MOVW and MOVT:
uint32_t diffw = diff & 0xffffu;
uint32_t difft = diff >> 16;
uint32_t movw = 0xf2400000u | // MOVW r0, #0 (placeholder),
((diffw & 0xf000u) << (16 - 12)) | // move imm4 from bits 12-15 to bits 16-19,
((diffw & 0x0800u) << (26 - 11)) | // move imm from bit 11 to bit 26,
((diffw & 0x0700u) << (12 - 8)) | // move imm3 from bits 8-10 to bits 12-14,
((diffw & 0x00ffu)); // keep imm8 at bits 0-7.
uint32_t movt = 0xf2c00000u | // MOVT r0, #0 (placeholder),
((difft & 0xf000u) << (16 - 12)) | // move imm4 from bits 12-15 to bits 16-19,
((difft & 0x0800u) << (26 - 11)) | // move imm from bit 11 to bit 26,
((difft & 0x0700u) << (12 - 8)) | // move imm3 from bits 8-10 to bits 12-14,
((difft & 0x00ffu)); // keep imm8 at bits 0-7.
const uint8_t expected_code[] = {
static_cast<uint8_t>(movw >> 16), static_cast<uint8_t>(movw >> 24),
static_cast<uint8_t>(movw >> 0), static_cast<uint8_t>(movw >> 8),
static_cast<uint8_t>(movt >> 16), static_cast<uint8_t>(movt >> 24),
static_cast<uint8_t>(movt >> 0), static_cast<uint8_t>(movt >> 8),
0x78, 0x44,
};
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallSelf) {
const LinkerPatch patches[] = {
LinkerPatch::RelativeCodePatch(0u, nullptr, 1u),
};
AddCompiledMethod(MethodRef(1u), kCallCode, ArrayRef<const LinkerPatch>(patches));
Link();
static const uint8_t expected_code[] = {
0xff, 0xf7, 0xfe, 0xff
};
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallOther) {
const LinkerPatch method1_patches[] = {
LinkerPatch::RelativeCodePatch(0u, nullptr, 2u),
};
AddCompiledMethod(MethodRef(1u), kCallCode, ArrayRef<const LinkerPatch>(method1_patches));
const LinkerPatch method2_patches[] = {
LinkerPatch::RelativeCodePatch(0u, nullptr, 1u),
};
AddCompiledMethod(MethodRef(2u), kCallCode, ArrayRef<const LinkerPatch>(method2_patches));
Link();
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t method2_offset = GetMethodOffset(2u);
uint32_t diff_after = method2_offset - (method1_offset + 4u /* PC adjustment */);
ASSERT_EQ(diff_after & 1u, 0u);
ASSERT_LT(diff_after >> 1, 1u << 8); // Simple encoding, (diff_after >> 1) fits into 8 bits.
static const uint8_t method1_expected_code[] = {
0x00, 0xf0, static_cast<uint8_t>(diff_after >> 1), 0xf8
};
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(method1_expected_code)));
uint32_t diff_before = method1_offset - (method2_offset + 4u /* PC adjustment */);
ASSERT_EQ(diff_before & 1u, 0u);
ASSERT_GE(diff_before, -1u << 9); // Simple encoding, -256 <= (diff >> 1) < 0.
auto method2_expected_code = GenNopsAndBl(0u, kBlMinus256 | ((diff_before >> 1) & 0xffu));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(2u), ArrayRef<const uint8_t>(method2_expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallTrampoline) {
const LinkerPatch patches[] = {
LinkerPatch::RelativeCodePatch(0u, nullptr, 2u),
};
AddCompiledMethod(MethodRef(1u), kCallCode, ArrayRef<const LinkerPatch>(patches));
Link();
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t diff = kTrampolineOffset - (method1_offset + 4u);
ASSERT_EQ(diff & 1u, 0u);
ASSERT_GE(diff, -1u << 9); // Simple encoding, -256 <= (diff >> 1) < 0 (checked as unsigned).
auto expected_code = GenNopsAndBl(0u, kBlMinus256 | ((diff >> 1) & 0xffu));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallTrampolineTooFar) {
constexpr uint32_t missing_method_index = 1024u;
auto last_method_raw_code = GenNopsAndBl(3u, kBlPlus0);
constexpr uint32_t bl_offset_in_last_method = 3u * 2u; // After NOPs.
ArrayRef<const uint8_t> last_method_code(last_method_raw_code);
ASSERT_EQ(bl_offset_in_last_method + 4u, last_method_code.size());
const LinkerPatch last_method_patches[] = {
LinkerPatch::RelativeCodePatch(bl_offset_in_last_method, nullptr, missing_method_index),
};
constexpr uint32_t just_over_max_negative_disp = 16 * MB + 2 - 4u /* PC adjustment */;
uint32_t last_method_idx = Create2MethodsWithGap(
kNopCode,
ArrayRef<const LinkerPatch>(),
last_method_code,
ArrayRef<const LinkerPatch>(last_method_patches),
just_over_max_negative_disp - bl_offset_in_last_method);
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t last_method_offset = GetMethodOffset(last_method_idx);
ASSERT_EQ(method1_offset,
last_method_offset + bl_offset_in_last_method - just_over_max_negative_disp);
ASSERT_FALSE(method_offset_map_.FindMethodOffset(MethodRef(missing_method_index)).first);
// Check linked code.
uint32_t thunk_offset = CompiledCode::AlignCode(
last_method_offset + last_method_code.size(), InstructionSet::kThumb2);
uint32_t diff =
thunk_offset - (last_method_offset + bl_offset_in_last_method + 4u /* PC adjustment */);
ASSERT_TRUE(IsAligned<2u>(diff));
ASSERT_LT(diff >> 1, 1u << 8); // Simple encoding, (diff >> 1) fits into 8 bits.
auto expected_code = GenNopsAndBl(3u, kBlPlus0 | ((diff >> 1) & 0xffu));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(last_method_idx),
ArrayRef<const uint8_t>(expected_code)));
EXPECT_TRUE(CheckThunk(thunk_offset));
}
TEST_F(Thumb2RelativePatcherTest, CallOtherAlmostTooFarAfter) {
auto method1_raw_code = GenNopsAndBl(3u, kBlPlus0);
constexpr uint32_t bl_offset_in_method1 = 3u * 2u; // After NOPs.
ArrayRef<const uint8_t> method1_code(method1_raw_code);
ASSERT_EQ(bl_offset_in_method1 + 4u, method1_code.size());
const uint32_t kExpectedLastMethodIdx = 9u; // Based on 2MiB chunks in Create2MethodsWithGap().
const LinkerPatch method1_patches[] = {
LinkerPatch::RelativeCodePatch(bl_offset_in_method1, nullptr, kExpectedLastMethodIdx),
};
constexpr uint32_t max_positive_disp = 16 * MB - 2u + 4u /* PC adjustment */;
uint32_t last_method_idx = Create2MethodsWithGap(method1_code,
ArrayRef<const LinkerPatch>(method1_patches),
kNopCode,
ArrayRef<const LinkerPatch>(),
bl_offset_in_method1 + max_positive_disp);
ASSERT_EQ(kExpectedLastMethodIdx, last_method_idx);
// Check linked code.
auto expected_code = GenNopsAndBl(3u, kBlPlusMax);
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallOtherAlmostTooFarBefore) {
auto last_method_raw_code = GenNopsAndBl(2u, kBlPlus0);
constexpr uint32_t bl_offset_in_last_method = 2u * 2u; // After NOPs.
ArrayRef<const uint8_t> last_method_code(last_method_raw_code);
ASSERT_EQ(bl_offset_in_last_method + 4u, last_method_code.size());
const LinkerPatch last_method_patches[] = {
LinkerPatch::RelativeCodePatch(bl_offset_in_last_method, nullptr, 1u),
};
constexpr uint32_t max_negative_disp = 16 * MB - 4u /* PC adjustment */;
uint32_t last_method_idx = Create2MethodsWithGap(kNopCode,
ArrayRef<const LinkerPatch>(),
last_method_code,
ArrayRef<const LinkerPatch>(last_method_patches),
max_negative_disp - bl_offset_in_last_method);
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t last_method_offset = GetMethodOffset(last_method_idx);
ASSERT_EQ(method1_offset, last_method_offset + bl_offset_in_last_method - max_negative_disp);
// Check linked code.
auto expected_code = GenNopsAndBl(2u, kBlMinusMax);
EXPECT_TRUE(CheckLinkedMethod(MethodRef(last_method_idx),
ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, CallOtherJustTooFarAfter) {
auto method1_raw_code = GenNopsAndBl(2u, kBlPlus0);
constexpr uint32_t bl_offset_in_method1 = 2u * 2u; // After NOPs.
ArrayRef<const uint8_t> method1_code(method1_raw_code);
ASSERT_EQ(bl_offset_in_method1 + 4u, method1_code.size());
const uint32_t kExpectedLastMethodIdx = 9u; // Based on 2MiB chunks in Create2MethodsWithGap().
const LinkerPatch method1_patches[] = {
LinkerPatch::RelativeCodePatch(bl_offset_in_method1, nullptr, kExpectedLastMethodIdx),
};
constexpr uint32_t just_over_max_positive_disp = 16 * MB + 4u /* PC adjustment */;
uint32_t last_method_idx = Create2MethodsWithGap(
method1_code,
ArrayRef<const LinkerPatch>(method1_patches),
kNopCode,
ArrayRef<const LinkerPatch>(),
bl_offset_in_method1 + just_over_max_positive_disp);
ASSERT_EQ(kExpectedLastMethodIdx, last_method_idx);
uint32_t method_after_thunk_idx = last_method_idx;
if (sizeof(OatQuickMethodHeader) < kArmCodeAlignment) {
// The thunk needs to start on a kArmCodeAlignment-aligned address before the address where the
// last method would have been if there was no thunk. If the size of the OatQuickMethodHeader
// is at least kArmCodeAlignment, the thunk start shall fit between the previous filler method
// and that address. Otherwise, it shall be inserted before that filler method.
method_after_thunk_idx -= 1u;
}
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t method_after_thunk_offset = GetMethodOffset(method_after_thunk_idx);
ASSERT_TRUE(IsAligned<kArmCodeAlignment>(method_after_thunk_offset));
uint32_t method_after_thunk_header_offset =
method_after_thunk_offset - sizeof(OatQuickMethodHeader);
uint32_t thunk_size = MethodCallThunkSize();
uint32_t thunk_offset =
RoundDown(method_after_thunk_header_offset - thunk_size, kArmCodeAlignment);
DCHECK_EQ(thunk_offset + thunk_size + CodeAlignmentSize(thunk_offset + thunk_size),
method_after_thunk_header_offset);
ASSERT_TRUE(IsAligned<kArmCodeAlignment>(thunk_offset));
uint32_t diff = thunk_offset - (method1_offset + bl_offset_in_method1 + 4u /* PC adjustment */);
ASSERT_TRUE(IsAligned<2u>(diff));
ASSERT_GE(diff, 16 * MB - (1u << 22)); // Simple encoding, unknown bits fit into imm10:imm11:0.
auto expected_code =
GenNopsAndBl(2u, 0xf000d000 | ((diff >> 1) & 0x7ffu) | (((diff >> 12) & 0x3ffu) << 16));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
CheckThunk(thunk_offset);
}
TEST_F(Thumb2RelativePatcherTest, CallOtherJustTooFarBefore) {
auto last_method_raw_code = GenNopsAndBl(3u, kBlPlus0);
constexpr uint32_t bl_offset_in_last_method = 3u * 2u; // After NOPs.
ArrayRef<const uint8_t> last_method_code(last_method_raw_code);
ASSERT_EQ(bl_offset_in_last_method + 4u, last_method_code.size());
const LinkerPatch last_method_patches[] = {
LinkerPatch::RelativeCodePatch(bl_offset_in_last_method, nullptr, 1u),
};
constexpr uint32_t just_over_max_negative_disp = 16 * MB + 2 - 4u /* PC adjustment */;
uint32_t last_method_idx = Create2MethodsWithGap(
kNopCode,
ArrayRef<const LinkerPatch>(),
last_method_code,
ArrayRef<const LinkerPatch>(last_method_patches),
just_over_max_negative_disp - bl_offset_in_last_method);
uint32_t method1_offset = GetMethodOffset(1u);
uint32_t last_method_offset = GetMethodOffset(last_method_idx);
ASSERT_EQ(method1_offset,
last_method_offset + bl_offset_in_last_method - just_over_max_negative_disp);
// Check linked code.
uint32_t thunk_offset = CompiledCode::AlignCode(
last_method_offset + last_method_code.size(), InstructionSet::kThumb2);
uint32_t diff =
thunk_offset - (last_method_offset + bl_offset_in_last_method + 4u /* PC adjustment */);
ASSERT_TRUE(IsAligned<2u>(diff));
ASSERT_LT(diff >> 1, 1u << 8); // Simple encoding, (diff >> 1) fits into 8 bits.
auto expected_code = GenNopsAndBl(3u, kBlPlus0 | ((diff >> 1) & 0xffu));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(last_method_idx),
ArrayRef<const uint8_t>(expected_code)));
EXPECT_TRUE(CheckThunk(thunk_offset));
}
TEST_F(Thumb2RelativePatcherTest, StringBssEntry1) {
TestStringBssEntry(0x00ff0000u, 0x00fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringBssEntry2) {
TestStringBssEntry(0x02ff0000u, 0x05fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringBssEntry3) {
TestStringBssEntry(0x08ff0000u, 0x08fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringBssEntry4) {
TestStringBssEntry(0xd0ff0000u, 0x60fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringReference1) {
TestStringReference(0x00ff00fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringReference2) {
TestStringReference(0x02ff05fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringReference3) {
TestStringReference(0x08ff08fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, StringReference4) {
TestStringReference(0xd0ff60fcu);
ASSERT_LT(GetMethodOffset(1u), 0xfcu);
}
TEST_F(Thumb2RelativePatcherTest, EntrypointCall) {
constexpr uint32_t kEntrypointOffset = 512;
const LinkerPatch patches[] = {
LinkerPatch::CallEntrypointPatch(0u, kEntrypointOffset),
};
AddCompiledMethod(MethodRef(1u), kCallCode, ArrayRef<const LinkerPatch>(patches));
Link();
uint32_t method_offset = GetMethodOffset(1u);
uint32_t thunk_offset = CompiledCode::AlignCode(method_offset + kCallCode.size(),
InstructionSet::kThumb2);
uint32_t diff = thunk_offset - method_offset - kPcAdjustment;
ASSERT_TRUE(IsAligned<2u>(diff));
ASSERT_LT(diff >> 1, 1u << 8); // Simple encoding, (diff >> 1) fits into 8 bits.
auto expected_code = GenNopsAndBl(0u, kBlPlus0 | ((diff >> 1) & 0xffu));
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
// Verify the thunk.
uint32_t ldr_pc_tr_offset =
0xf8d00000 | // LDR Rt, [Rn, #<imm12>]
(/* tr */ 9 << 16) | // Rn = TR
(/* pc */ 15 << 12) | // Rt = PC
kEntrypointOffset; // imm12
uint16_t bkpt = 0xbe00;
ASSERT_LE(6u, output_.size() - thunk_offset);
EXPECT_EQ(ldr_pc_tr_offset, GetOutputInsn32(thunk_offset));
EXPECT_EQ(bkpt, GetOutputInsn16(thunk_offset + 4u));
}
const uint32_t kBakerValidRegs[] = {
0, 1, 2, 3, 4, 5, 6, 7,
9, 10, 11, // r8 (rMR), IP, SP, LR and PC are reserved.
};
const uint32_t kBakerValidRegsNarrow[] = {
0, 1, 2, 3, 4, 5, 6, 7,
};
void Thumb2RelativePatcherTest::TestBakerFieldWide(uint32_t offset,
uint32_t ref_reg,
bool implicit_null_checks) {
implicit_null_checks_ = implicit_null_checks;
DCHECK_ALIGNED(offset, 4u);
DCHECK_LT(offset, 4 * KB);
constexpr size_t kMethodCodeSize = 8u;
constexpr size_t kLiteralOffset = 0u;
uint32_t method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
for (uint32_t holder_reg : kBakerValidRegs) {
uint32_t ldr = kLdrWInsn | offset | (base_reg << 16) | (ref_reg << 12);
const std::vector<uint8_t> raw_code = RawCode({kBneWPlus0, ldr});
ASSERT_EQ(kMethodCodeSize, raw_code.size());
ArrayRef<const uint8_t> code(raw_code);
uint32_t encoded_data = EncodeBakerReadBarrierFieldData(
base_reg, holder_reg, /* narrow */ false);
const LinkerPatch patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset, encoded_data),
};
++method_idx;
AddCompiledMethod(MethodRef(method_idx), code, ArrayRef<const LinkerPatch>(patches));
}
}
Link();
// All thunks are at the end.
uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment);
method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
for (uint32_t holder_reg : kBakerValidRegs) {
++method_idx;
uint32_t bne = BneWWithOffset(GetMethodOffset(method_idx) + kLiteralOffset, thunk_offset);
uint32_t ldr = kLdrWInsn | offset | (base_reg << 16) | (ref_reg << 12);
const std::vector<uint8_t> expected_code = RawCode({bne, ldr});
ASSERT_EQ(kMethodCodeSize, expected_code.size()) << "bne=0x" << std::hex << bne;
ASSERT_TRUE(
CheckLinkedMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(expected_code)));
std::vector<uint8_t> expected_thunk =
CompileBakerOffsetThunk(base_reg, holder_reg, /* narrow */ false);
ASSERT_GT(output_.size(), thunk_offset);
ASSERT_GE(output_.size() - thunk_offset, expected_thunk.size());
ArrayRef<const uint8_t> compiled_thunk(output_.data() + thunk_offset,
expected_thunk.size());
if (ArrayRef<const uint8_t>(expected_thunk) != compiled_thunk) {
DumpDiff(ArrayRef<const uint8_t>(expected_thunk), compiled_thunk);
ASSERT_TRUE(false);
}
size_t gray_check_offset = thunk_offset;
if (implicit_null_checks && holder_reg == base_reg) {
// Verify that the null-check uses the correct register, i.e. holder_reg.
if (holder_reg < 8) {
ASSERT_GE(output_.size() - gray_check_offset, 2u);
ASSERT_EQ(0xb100 | holder_reg, GetOutputInsn16(thunk_offset) & 0xfd07u);
gray_check_offset +=2u;
} else {
ASSERT_GE(output_.size() - gray_check_offset, 6u);
ASSERT_EQ(0xf1b00f00u | (holder_reg << 16), GetOutputInsn32(thunk_offset) & 0xfbff8f00u);
ASSERT_EQ(0xd000u, GetOutputInsn16(thunk_offset + 4u) & 0xff00u); // BEQ
gray_check_offset += 6u;
}
}
// Verify that the lock word for gray bit check is loaded from the holder address.
ASSERT_GE(output_.size() - gray_check_offset,
4u * /* 32-bit instructions */ 4u + 2u * /* 16-bit instructions */ 2u);
const uint32_t load_lock_word =
kLdrWInsn |
(holder_reg << 16) |
(/* IP */ 12 << 12) |
mirror::Object::MonitorOffset().Uint32Value();
ASSERT_EQ(load_lock_word, GetOutputInsn32(gray_check_offset));
// Verify the gray bit check.
DCHECK_GE(LockWord::kReadBarrierStateShift, 8u); // ROR modified immediate.
uint32_t ror_shift = 7 + (32 - LockWord::kReadBarrierStateShift);
const uint32_t tst_gray_bit_without_offset =
0xf0100f00 | (/* IP */ 12 << 16)
| (((ror_shift >> 4) & 1) << 26) // i
| (((ror_shift >> 1) & 7) << 12) // imm3
| ((ror_shift & 1) << 7); // imm8, ROR('1':imm8<7:0>, ror_shift).
EXPECT_EQ(tst_gray_bit_without_offset, GetOutputInsn32(gray_check_offset + 4u));
EXPECT_EQ(0xd100u, GetOutputInsn16(gray_check_offset + 8u) & 0xff00u); // BNE
// Verify the fake dependency (skip "ADD LR, LR, #ldr_offset").
const uint32_t fake_dependency =
0xeb000010 | // ADD Rd, Rn, Rm, LSR 32 (type=01, imm3=000, imm2=00)
(/* IP */ 12) | // Rm = IP
(base_reg << 16) | // Rn = base_reg
(base_reg << 8); // Rd = base_reg
EXPECT_EQ(fake_dependency, GetOutputInsn32(gray_check_offset + 14u));
// Do not check the rest of the implementation.
// The next thunk follows on the next aligned offset.
thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment);
}
}
}
void Thumb2RelativePatcherTest::TestBakerFieldNarrow(uint32_t offset,
uint32_t ref_reg,
bool implicit_null_checks) {
implicit_null_checks_ = implicit_null_checks;
DCHECK_ALIGNED(offset, 4u);
DCHECK_LT(offset, 32u);
constexpr size_t kMethodCodeSize = 6u;
constexpr size_t kLiteralOffset = 0u;
uint32_t method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
if (base_reg >= 8u) {
continue;
}
for (uint32_t holder_reg : kBakerValidRegs) {
uint32_t ldr = kLdrInsn | (offset << (6 - 2)) | (base_reg << 3) | ref_reg;
const std::vector<uint8_t> raw_code = RawCode({kBneWPlus0, ldr});
ASSERT_EQ(kMethodCodeSize, raw_code.size());
ArrayRef<const uint8_t> code(raw_code);
uint32_t encoded_data = EncodeBakerReadBarrierFieldData(
base_reg, holder_reg, /* narrow */ true);
const LinkerPatch patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset, encoded_data),
};
++method_idx;
AddCompiledMethod(MethodRef(method_idx), code, ArrayRef<const LinkerPatch>(patches));
}
}
Link();
// All thunks are at the end.
uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment);
method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
if (base_reg >= 8u) {
continue;
}
for (uint32_t holder_reg : kBakerValidRegs) {
++method_idx;
uint32_t bne = BneWWithOffset(GetMethodOffset(method_idx) + kLiteralOffset, thunk_offset);
uint32_t ldr = kLdrInsn | (offset << (6 - 2)) | (base_reg << 3) | ref_reg;
const std::vector<uint8_t> expected_code = RawCode({bne, ldr});
ASSERT_EQ(kMethodCodeSize, expected_code.size()) << "bne=0x" << std::hex << bne;
ASSERT_TRUE(
CheckLinkedMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(expected_code)));
std::vector<uint8_t> expected_thunk =
CompileBakerOffsetThunk(base_reg, holder_reg, /* narrow */ true);
ASSERT_GT(output_.size(), thunk_offset);
ASSERT_GE(output_.size() - thunk_offset, expected_thunk.size());
ArrayRef<const uint8_t> compiled_thunk(output_.data() + thunk_offset,
expected_thunk.size());
if (ArrayRef<const uint8_t>(expected_thunk) != compiled_thunk) {
DumpDiff(ArrayRef<const uint8_t>(expected_thunk), compiled_thunk);
ASSERT_TRUE(false);
}
size_t gray_check_offset = thunk_offset;
if (implicit_null_checks && holder_reg == base_reg) {
// Verify that the null-check uses the correct register, i.e. holder_reg.
if (holder_reg < 8) {
ASSERT_GE(output_.size() - gray_check_offset, 2u);
ASSERT_EQ(0xb100 | holder_reg, GetOutputInsn16(thunk_offset) & 0xfd07u);
gray_check_offset +=2u;
} else {
ASSERT_GE(output_.size() - gray_check_offset, 6u);
ASSERT_EQ(0xf1b00f00u | (holder_reg << 16), GetOutputInsn32(thunk_offset) & 0xfbff8f00u);
ASSERT_EQ(0xd000u, GetOutputInsn16(thunk_offset + 4u) & 0xff00u); // BEQ
gray_check_offset += 6u;
}
}
// Verify that the lock word for gray bit check is loaded from the holder address.
ASSERT_GE(output_.size() - gray_check_offset,
4u * /* 32-bit instructions */ 4u + 2u * /* 16-bit instructions */ 2u);
const uint32_t load_lock_word =
kLdrWInsn |
(holder_reg << 16) |
(/* IP */ 12 << 12) |
mirror::Object::MonitorOffset().Uint32Value();
ASSERT_EQ(load_lock_word, GetOutputInsn32(gray_check_offset));
// Verify the gray bit check.
DCHECK_GE(LockWord::kReadBarrierStateShift, 8u); // ROR modified immediate.
uint32_t ror_shift = 7 + (32 - LockWord::kReadBarrierStateShift);
const uint32_t tst_gray_bit_without_offset =
0xf0100f00 | (/* IP */ 12 << 16)
| (((ror_shift >> 4) & 1) << 26) // i
| (((ror_shift >> 1) & 7) << 12) // imm3
| ((ror_shift & 1) << 7); // imm8, ROR('1':imm8<7:0>, ror_shift).
EXPECT_EQ(tst_gray_bit_without_offset, GetOutputInsn32(gray_check_offset + 4u));
EXPECT_EQ(0xd100u, GetOutputInsn16(gray_check_offset + 8u) & 0xff00u); // BNE
// Verify the fake dependency (skip "ADD LR, LR, #ldr_offset").
const uint32_t fake_dependency =
0xeb000010 | // ADD Rd, Rn, Rm, LSR 32 (type=01, imm3=000, imm2=00)
(/* IP */ 12) | // Rm = IP
(base_reg << 16) | // Rn = base_reg
(base_reg << 8); // Rd = base_reg
EXPECT_EQ(fake_dependency, GetOutputInsn32(gray_check_offset + 14u));
// Do not check the rest of the implementation.
// The next thunk follows on the next aligned offset.
thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment);
}
}
}
TEST_F(Thumb2RelativePatcherTest, BakerOffsetWide) {
struct TestCase {
uint32_t offset;
uint32_t ref_reg;
};
static const TestCase test_cases[] = {
{ 0u, 0u },
{ 8u, 3u },
{ 28u, 7u },
{ 0xffcu, 11u },
};
for (const TestCase& test_case : test_cases) {
Reset();
TestBakerFieldWide(test_case.offset, test_case.ref_reg, /*implicit_null_checks=*/ true);
Reset();
TestBakerFieldWide(test_case.offset, test_case.ref_reg, /*implicit_null_checks=*/ false);
}
}
TEST_F(Thumb2RelativePatcherTest, BakerOffsetNarrow) {
struct TestCase {
uint32_t offset;
uint32_t ref_reg;
};
static const TestCase test_cases[] = {
{ 0, 0u },
{ 8, 3u },
{ 28, 7u },
};
for (const TestCase& test_case : test_cases) {
Reset();
TestBakerFieldNarrow(test_case.offset, test_case.ref_reg, /*implicit_null_checks=*/ true);
Reset();
TestBakerFieldNarrow(test_case.offset, test_case.ref_reg, /*implicit_null_checks=*/ false);
}
}
TEST_F(Thumb2RelativePatcherTest, BakerOffsetThunkInTheMiddle) {
// One thunk in the middle with maximum distance branches to it from both sides.
// Use offset = 0, base_reg = 0, ref_reg = 0, the LDR is simply `kLdrWInsn`.
constexpr uint32_t kLiteralOffset1 = 6u;
const std::vector<uint8_t> raw_code1 = RawCode({kNopWInsn, kNopInsn, kBneWPlus0, kLdrWInsn});
ArrayRef<const uint8_t> code1(raw_code1);
uint32_t encoded_data = EncodeBakerReadBarrierFieldData(
/* base_reg */ 0, /* holder_reg */ 0, /* narrow */ false);
const LinkerPatch patches1[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset1, encoded_data),
};
AddCompiledMethod(MethodRef(1u), code1, ArrayRef<const LinkerPatch>(patches1));
constexpr uint32_t expected_thunk_offset =
kLiteralOffset1 + kPcAdjustment + /* kMaxBcondPositiveDisplacement */ ((1 << 20) - 2u);
static_assert(IsAligned<kArmCodeAlignment>(expected_thunk_offset),
"Target offset must be aligned.");
size_t filler1_size = expected_thunk_offset -
RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment);
std::vector<uint8_t> raw_filler1_code = GenNops(filler1_size / 2u);
ArrayRef<const uint8_t> filler1_code(raw_filler1_code);
AddCompiledMethod(MethodRef(2u), filler1_code);
// Enforce thunk reservation with a tiny method.
AddCompiledMethod(MethodRef(3u), kNopCode);
constexpr uint32_t kLiteralOffset2 = 4;
static_assert(IsAligned<kArmCodeAlignment>(kLiteralOffset2 + kPcAdjustment),
"PC for BNE must be aligned.");
// Allow reaching the thunk from the very beginning of a method almost 1MiB away. Backward branch
// reaches the full 1MiB but we need to take PC adjustment into account. Things to subtract:
// - thunk size and method 3 pre-header, rounded up (padding in between if needed)
// - method 3 code and method 4 pre-header, rounded up (padding in between if needed)
// - method 4 header (let there be no padding between method 4 code and method 5 pre-header).
size_t thunk_size =
CompileBakerOffsetThunk(/* base_reg */ 0, /* holder_reg */ 0, /* narrow */ false).size();
size_t filler2_size =
1 * MB - (kLiteralOffset2 + kPcAdjustment)
- RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmCodeAlignment)
- RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment)
- sizeof(OatQuickMethodHeader);
std::vector<uint8_t> raw_filler2_code = GenNops(filler2_size / 2u);
ArrayRef<const uint8_t> filler2_code(raw_filler2_code);
AddCompiledMethod(MethodRef(4u), filler2_code);
const std::vector<uint8_t> raw_code2 = RawCode({kNopWInsn, kBneWPlus0, kLdrWInsn});
ArrayRef<const uint8_t> code2(raw_code2);
const LinkerPatch patches2[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset2, encoded_data),
};
AddCompiledMethod(MethodRef(5u), code2, ArrayRef<const LinkerPatch>(patches2));
Link();
uint32_t first_method_offset = GetMethodOffset(1u);
uint32_t last_method_offset = GetMethodOffset(5u);
EXPECT_EQ(2 * MB, last_method_offset - first_method_offset);
const uint32_t bne_max_forward = kBneWPlus0 | 0x003f2fff;
const uint32_t bne_max_backward = kBneWPlus0 | 0x04000000;
const std::vector<uint8_t> expected_code1 =
RawCode({kNopWInsn, kNopInsn, bne_max_forward, kLdrWInsn});
const std::vector<uint8_t> expected_code2 = RawCode({kNopWInsn, bne_max_backward, kLdrWInsn});
ASSERT_TRUE(CheckLinkedMethod(MethodRef(1), ArrayRef<const uint8_t>(expected_code1)));
ASSERT_TRUE(CheckLinkedMethod(MethodRef(5), ArrayRef<const uint8_t>(expected_code2)));
}
TEST_F(Thumb2RelativePatcherTest, BakerOffsetThunkBeforeFiller) {
// Based on the first part of BakerOffsetThunkInTheMiddle but the BNE is one instruction
// earlier, so the thunk is emitted before the filler.
// Use offset = 0, base_reg = 0, ref_reg = 0, the LDR is simply `kLdrWInsn`.
constexpr uint32_t kLiteralOffset1 = 4u;
const std::vector<uint8_t> raw_code1 = RawCode({kNopWInsn, kBneWPlus0, kLdrWInsn, kNopInsn});
ArrayRef<const uint8_t> code1(raw_code1);
uint32_t encoded_data = EncodeBakerReadBarrierFieldData(
/* base_reg */ 0, /* holder_reg */ 0, /* narrow */ false);
const LinkerPatch patches1[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset1, encoded_data),
};
AddCompiledMethod(MethodRef(1u), code1, ArrayRef<const LinkerPatch>(patches1));
constexpr uint32_t expected_thunk_offset =
kLiteralOffset1 + kPcAdjustment + /* kMaxBcondPositiveDisplacement + 2 */ (1u << 20);
static_assert(IsAligned<kArmCodeAlignment>(expected_thunk_offset),
"Target offset must be aligned.");
size_t filler1_size = expected_thunk_offset -
RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment);
std::vector<uint8_t> raw_filler1_code = GenNops(filler1_size / 2u);
ArrayRef<const uint8_t> filler1_code(raw_filler1_code);
AddCompiledMethod(MethodRef(2u), filler1_code);
Link();
const uint32_t bne =
BneWWithOffset(kLiteralOffset1, RoundUp(raw_code1.size(), kArmCodeAlignment));
const std::vector<uint8_t> expected_code1 = RawCode({kNopWInsn, bne, kLdrWInsn, kNopInsn});
ASSERT_TRUE(CheckLinkedMethod(MethodRef(1), ArrayRef<const uint8_t>(expected_code1)));
}
TEST_F(Thumb2RelativePatcherTest, BakerOffsetThunkInTheMiddleUnreachableFromLast) {
// Based on the BakerOffsetThunkInTheMiddle but the BNE in the last method is preceded
// by NOP and cannot reach the thunk in the middle, so we emit an extra thunk at the end.
// Use offset = 0, base_reg = 0, ref_reg = 0, the LDR is simply `kLdrWInsn`.
constexpr uint32_t kLiteralOffset1 = 6u;
const std::vector<uint8_t> raw_code1 = RawCode({kNopWInsn, kNopInsn, kBneWPlus0, kLdrWInsn});
ArrayRef<const uint8_t> code1(raw_code1);
uint32_t encoded_data = EncodeBakerReadBarrierFieldData(
/* base_reg */ 0, /* holder_reg */ 0, /* narrow */ false);
const LinkerPatch patches1[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset1, encoded_data),
};
AddCompiledMethod(MethodRef(1u), code1, ArrayRef<const LinkerPatch>(patches1));
constexpr uint32_t expected_thunk_offset =
kLiteralOffset1 + kPcAdjustment + /* kMaxBcondPositiveDisplacement */ ((1 << 20) - 2u);
static_assert(IsAligned<kArmCodeAlignment>(expected_thunk_offset),
"Target offset must be aligned.");
size_t filler1_size = expected_thunk_offset -
RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment);
std::vector<uint8_t> raw_filler1_code = GenNops(filler1_size / 2u);
ArrayRef<const uint8_t> filler1_code(raw_filler1_code);
AddCompiledMethod(MethodRef(2u), filler1_code);
// Enforce thunk reservation with a tiny method.
AddCompiledMethod(MethodRef(3u), kNopCode);
constexpr uint32_t kReachableFromOffset2 = 4;
constexpr uint32_t kLiteralOffset2 = kReachableFromOffset2 + 2;
static_assert(IsAligned<kArmCodeAlignment>(kReachableFromOffset2 + kPcAdjustment),
"PC for BNE must be aligned.");
// If not for the extra NOP, this would allow reaching the thunk from the BNE
// of a method 1MiB away. Backward branch reaches the full 1MiB but we need to take
// PC adjustment into account. Things to subtract:
// - thunk size and method 3 pre-header, rounded up (padding in between if needed)
// - method 3 code and method 4 pre-header, rounded up (padding in between if needed)
// - method 4 header (let there be no padding between method 4 code and method 5 pre-header).
size_t thunk_size =
CompileBakerOffsetThunk(/* base_reg */ 0, /* holder_reg */ 0, /* narrow */ false).size();
size_t filler2_size =
1 * MB - (kReachableFromOffset2 + kPcAdjustment)
- RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmCodeAlignment)
- RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment)
- sizeof(OatQuickMethodHeader);
std::vector<uint8_t> raw_filler2_code = GenNops(filler2_size / 2u);
ArrayRef<const uint8_t> filler2_code(raw_filler2_code);
AddCompiledMethod(MethodRef(4u), filler2_code);
// Extra 16-bit NOP compared to BakerOffsetThunkInTheMiddle.
const std::vector<uint8_t> raw_code2 = RawCode({kNopWInsn, kNopInsn, kBneWPlus0, kLdrWInsn});
ArrayRef<const uint8_t> code2(raw_code2);
const LinkerPatch patches2[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kLiteralOffset2, encoded_data),
};
AddCompiledMethod(MethodRef(5u), code2, ArrayRef<const LinkerPatch>(patches2));
Link();
uint32_t first_method_offset = GetMethodOffset(1u);
uint32_t last_method_offset = GetMethodOffset(5u);
EXPECT_EQ(2 * MB, last_method_offset - first_method_offset);
const uint32_t bne_max_forward = kBneWPlus0 | 0x003f2fff;
const uint32_t bne_last =
BneWWithOffset(kLiteralOffset2, RoundUp(raw_code2.size(), kArmCodeAlignment));
const std::vector<uint8_t> expected_code1 =
RawCode({kNopWInsn, kNopInsn, bne_max_forward, kLdrWInsn});
const std::vector<uint8_t> expected_code2 =
RawCode({kNopWInsn, kNopInsn, bne_last, kLdrWInsn});
ASSERT_TRUE(CheckLinkedMethod(MethodRef(1), ArrayRef<const uint8_t>(expected_code1)));
ASSERT_TRUE(CheckLinkedMethod(MethodRef(5), ArrayRef<const uint8_t>(expected_code2)));
}
TEST_F(Thumb2RelativePatcherTest, BakerArray) {
auto ldr = [](uint32_t base_reg) {
uint32_t index_reg = (base_reg == 0u) ? 1u : 0u;
uint32_t ref_reg = (base_reg == 2) ? 3u : 2u;
return kLdrRegLsl2 | index_reg | (base_reg << 16) | (ref_reg << 12);
};
constexpr size_t kMethodCodeSize = 8u;
constexpr size_t kLiteralOffset = 0u;
uint32_t method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
++method_idx;
const std::vector<uint8_t> raw_code = RawCode({kBneWPlus0, ldr(base_reg)});
ASSERT_EQ(kMethodCodeSize, raw_code.size());
ArrayRef<const uint8_t> code(raw_code);
const LinkerPatch patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(
kLiteralOffset, EncodeBakerReadBarrierArrayData(base_reg)),
};
AddCompiledMethod(MethodRef(method_idx), code, ArrayRef<const LinkerPatch>(patches));
}
Link();
// All thunks are at the end.
uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment);
method_idx = 0u;
for (uint32_t base_reg : kBakerValidRegs) {
++method_idx;
uint32_t bne = BneWWithOffset(GetMethodOffset(method_idx) + kLiteralOffset, thunk_offset);
const std::vector<uint8_t> expected_code = RawCode({bne, ldr(base_reg)});
ASSERT_EQ(kMethodCodeSize, expected_code.size());
EXPECT_TRUE(CheckLinkedMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(expected_code)));
std::vector<uint8_t> expected_thunk = CompileBakerArrayThunk(base_reg);
ASSERT_GT(output_.size(), thunk_offset);
ASSERT_GE(output_.size() - thunk_offset, expected_thunk.size());
ArrayRef<const uint8_t> compiled_thunk(output_.data() + thunk_offset,
expected_thunk.size());
if (ArrayRef<const uint8_t>(expected_thunk) != compiled_thunk) {
DumpDiff(ArrayRef<const uint8_t>(expected_thunk), compiled_thunk);
ASSERT_TRUE(false);
}
// Verify that the lock word for gray bit check is loaded from the correct address
// before the base_reg which points to the array data.
ASSERT_GE(output_.size() - thunk_offset,
4u * /* 32-bit instructions */ 4u + 2u * /* 16-bit instructions */ 2u);
int32_t data_offset =
mirror::Array::DataOffset(Primitive::ComponentSize(Primitive::kPrimNot)).Int32Value();
int32_t offset = mirror::Object::MonitorOffset().Int32Value() - data_offset;
ASSERT_LT(offset, 0);
ASSERT_GT(offset, -256);
const uint32_t load_lock_word =
kLdrNegativeOffset |
(-offset & 0xffu) |
(base_reg << 16) |
(/* IP */ 12 << 12);
EXPECT_EQ(load_lock_word, GetOutputInsn32(thunk_offset));
// Verify the gray bit check.
DCHECK_GE(LockWord::kReadBarrierStateShift, 8u); // ROR modified immediate.
uint32_t ror_shift = 7 + (32 - LockWord::kReadBarrierStateShift);
const uint32_t tst_gray_bit_without_offset =
0xf0100f00 | (/* IP */ 12 << 16)
| (((ror_shift >> 4) & 1) << 26) // i
| (((ror_shift >> 1) & 7) << 12) // imm3
| ((ror_shift & 1) << 7); // imm8, ROR('1':imm8<7:0>, ror_shift).
EXPECT_EQ(tst_gray_bit_without_offset, GetOutputInsn32(thunk_offset + 4u));
EXPECT_EQ(0xd100u, GetOutputInsn16(thunk_offset + 8u) & 0xff00u); // BNE
// Verify the fake dependency.
const uint32_t fake_dependency =
0xeb000010 | // ADD Rd, Rn, Rm, LSR 32 (type=01, imm3=000, imm2=00)
(/* IP */ 12) | // Rm = IP
(base_reg << 16) | // Rn = base_reg
(base_reg << 8); // Rd = base_reg
EXPECT_EQ(fake_dependency, GetOutputInsn32(thunk_offset + 14u));
// Do not check the rest of the implementation.
// The next thunk follows on the next aligned offset.
thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment);
}
}
TEST_F(Thumb2RelativePatcherTest, BakerGcRootWide) {
constexpr size_t kMethodCodeSize = 8u;
constexpr size_t kLiteralOffset = 4u;
uint32_t method_idx = 0u;
for (uint32_t root_reg : kBakerValidRegs) {
++method_idx;
uint32_t ldr = kLdrWInsn | (/* offset */ 8) | (/* base_reg */ 0 << 16) | (root_reg << 12);
const std::vector<uint8_t> raw_code = RawCode({ldr, kBneWPlus0});
ASSERT_EQ(kMethodCodeSize, raw_code.size());
ArrayRef<const uint8_t> code(raw_code);
const LinkerPatch patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(
kLiteralOffset, EncodeBakerReadBarrierGcRootData(root_reg, /* narrow */ false)),
};
AddCompiledMethod(MethodRef(method_idx), code, ArrayRef<const LinkerPatch>(patches));
}
Link();
// All thunks are at the end.
uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment);
method_idx = 0u;
for (uint32_t root_reg : kBakerValidRegs) {
++method_idx;
uint32_t bne = BneWWithOffset(GetMethodOffset(method_idx) + kLiteralOffset, thunk_offset);
uint32_t ldr = kLdrWInsn | (/* offset */ 8) | (/* base_reg */ 0 << 16) | (root_reg << 12);
const std::vector<uint8_t> expected_code = RawCode({ldr, bne});
ASSERT_EQ(kMethodCodeSize, expected_code.size());
EXPECT_TRUE(CheckLinkedMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(expected_code)));
std::vector<uint8_t> expected_thunk = CompileBakerGcRootThunk(root_reg, /* narrow */ false);
ASSERT_GT(output_.size(), thunk_offset);
ASSERT_GE(output_.size() - thunk_offset, expected_thunk.size());
ArrayRef<const uint8_t> compiled_thunk(output_.data() + thunk_offset,
expected_thunk.size());
if (ArrayRef<const uint8_t>(expected_thunk) != compiled_thunk) {
DumpDiff(ArrayRef<const uint8_t>(expected_thunk), compiled_thunk);
ASSERT_TRUE(false);
}
// Verify that the fast-path null-check uses the correct register, i.e. root_reg.
if (root_reg < 8) {
ASSERT_GE(output_.size() - thunk_offset, 2u);
ASSERT_EQ(0xb100 | root_reg, GetOutputInsn16(thunk_offset) & 0xfd07u);
} else {
ASSERT_GE(output_.size() - thunk_offset, 6u);
ASSERT_EQ(0xf1b00f00u | (root_reg << 16), GetOutputInsn32(thunk_offset) & 0xfbff8f00u);
ASSERT_EQ(0xd000u, GetOutputInsn16(thunk_offset + 4u) & 0xff00u); // BEQ
}
// Do not check the rest of the implementation.
// The next thunk follows on the next aligned offset.
thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment);
}
}
TEST_F(Thumb2RelativePatcherTest, BakerGcRootNarrow) {
constexpr size_t kMethodCodeSize = 6u;
constexpr size_t kLiteralOffset = 2u;
uint32_t method_idx = 0u;
for (uint32_t root_reg : kBakerValidRegsNarrow) {
++method_idx;
uint32_t ldr = kLdrInsn | (/* offset */ 8 << (6 - 2)) | (/* base_reg */ 0 << 3) | root_reg;
const std::vector<uint8_t> raw_code = RawCode({ldr, kBneWPlus0});
ASSERT_EQ(kMethodCodeSize, raw_code.size());
ArrayRef<const uint8_t> code(raw_code);
const LinkerPatch patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(
kLiteralOffset, EncodeBakerReadBarrierGcRootData(root_reg, /* narrow */ true)),
};
AddCompiledMethod(MethodRef(method_idx), code, ArrayRef<const LinkerPatch>(patches));
}
Link();
// All thunks are at the end.
uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment);
method_idx = 0u;
for (uint32_t root_reg : kBakerValidRegsNarrow) {
++method_idx;
uint32_t bne = BneWWithOffset(GetMethodOffset(method_idx) + kLiteralOffset, thunk_offset);
uint32_t ldr = kLdrInsn | (/* offset */ 8 << (6 - 2)) | (/* base_reg */ 0 << 3) | root_reg;
const std::vector<uint8_t> expected_code = RawCode({ldr, bne});
ASSERT_EQ(kMethodCodeSize, expected_code.size());
EXPECT_TRUE(CheckLinkedMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(expected_code)));
std::vector<uint8_t> expected_thunk = CompileBakerGcRootThunk(root_reg, /* narrow */ true);
ASSERT_GT(output_.size(), thunk_offset);
ASSERT_GE(output_.size() - thunk_offset, expected_thunk.size());
ArrayRef<const uint8_t> compiled_thunk(output_.data() + thunk_offset,
expected_thunk.size());
if (ArrayRef<const uint8_t>(expected_thunk) != compiled_thunk) {
DumpDiff(ArrayRef<const uint8_t>(expected_thunk), compiled_thunk);
ASSERT_TRUE(false);
}
// Verify that the fast-path null-check CBZ uses the correct register, i.e. root_reg.
ASSERT_GE(output_.size() - thunk_offset, 2u);
ASSERT_EQ(0xb100 | root_reg, GetOutputInsn16(thunk_offset) & 0xfd07u);
// Do not check the rest of the implementation.
// The next thunk follows on the next aligned offset.
thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment);
}
}
TEST_F(Thumb2RelativePatcherTest, BakerGcRootOffsetBits) {
// Test 1MiB of patches to the same thunk to stress-test different large offsets.
// (The low bits are not that important but the location of the high bits is easy to get wrong.)
std::vector<uint8_t> code;
code.reserve(1 * MB);
const size_t num_patches = 1 * MB / 8u;
std::vector<LinkerPatch> patches;
patches.reserve(num_patches);
const uint32_t ldr =
kLdrWInsn | (/* offset */ 8) | (/* base_reg */ 0 << 16) | (/* root_reg */ 0 << 12);
uint32_t encoded_data = EncodeBakerReadBarrierGcRootData(/* root_reg */ 0, /* narrow */ false);
for (size_t i = 0; i != num_patches; ++i) {
PushBackInsn(&code, ldr);
PushBackInsn(&code, kBneWPlus0);
patches.push_back(LinkerPatch::BakerReadBarrierBranchPatch(8u * i + 4u, encoded_data));
}
ASSERT_EQ(1 * MB, code.size());
ASSERT_EQ(num_patches, patches.size());
AddCompiledMethod(MethodRef(1u),
ArrayRef<const uint8_t>(code),
ArrayRef<const LinkerPatch>(patches));
Link();
// The thunk is right after the method code.
DCHECK_ALIGNED(1 * MB, kArmCodeAlignment);
std::vector<uint8_t> expected_code;
for (size_t i = 0; i != num_patches; ++i) {
PushBackInsn(&expected_code, ldr);
PushBackInsn(&expected_code, BneWWithOffset(8u * i + 4u, 1 * MB));
patches.push_back(LinkerPatch::BakerReadBarrierBranchPatch(8u * i + 4u, encoded_data));
}
EXPECT_TRUE(CheckLinkedMethod(MethodRef(1u), ArrayRef<const uint8_t>(expected_code)));
}
TEST_F(Thumb2RelativePatcherTest, BakerAndMethodCallInteraction) {
// During development, there was a `DCHECK_LE(MaxNextOffset(), next_thunk.MaxNextOffset());`
// in `ArmBaseRelativePatcher::ThunkData::MakeSpaceBefore()` which does not necessarily
// hold when we're reserving thunks of different sizes. This test exposes the situation
// by using Baker thunks and a method call thunk.
// Add a method call patch that can reach to method 1 offset + 16MiB.
uint32_t method_idx = 0u;
constexpr size_t kMethodCallLiteralOffset = 2u;
constexpr uint32_t kMissingMethodIdx = 2u;
const std::vector<uint8_t> raw_code1 = RawCode({kNopInsn, kBlPlus0});
const LinkerPatch method1_patches[] = {
LinkerPatch::RelativeCodePatch(kMethodCallLiteralOffset, nullptr, 2u),
};
ArrayRef<const uint8_t> code1(raw_code1);
++method_idx;
AddCompiledMethod(MethodRef(1u), code1, ArrayRef<const LinkerPatch>(method1_patches));
// Skip kMissingMethodIdx.
++method_idx;
ASSERT_EQ(kMissingMethodIdx, method_idx);
// Add a method with the right size that the method code for the next one starts 1MiB
// after code for method 1.
size_t filler_size =
1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment)
- sizeof(OatQuickMethodHeader);
std::vector<uint8_t> filler_code = GenNops(filler_size / 2u);
++method_idx;
AddCompiledMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(filler_code));
// Add 14 methods with 1MiB code+header, making the code for the next method start 1MiB
// before the currently scheduled MaxNextOffset() for the method call thunk.
for (uint32_t i = 0; i != 14; ++i) {
filler_size = 1 * MB - sizeof(OatQuickMethodHeader);
filler_code = GenNops(filler_size / 2u);
++method_idx;
AddCompiledMethod(MethodRef(method_idx), ArrayRef<const uint8_t>(filler_code));
}
// Add 2 Baker GC root patches to the last method, one that would allow the thunk at
// 1MiB + kArmCodeAlignment, i.e. kArmCodeAlignment after the method call thunk, and the
// second that needs it kArmCodeAlignment after that. Given the size of the GC root thunk
// is more than the space required by the method call thunk plus kArmCodeAlignment,
// this pushes the first GC root thunk's pending MaxNextOffset() before the method call
// thunk's pending MaxNextOffset() which needs to be adjusted.
ASSERT_LT(RoundUp(CompileMethodCallThunk().size(), kArmCodeAlignment) + kArmCodeAlignment,
CompileBakerGcRootThunk(/* root_reg */ 0, /* narrow */ false).size());
static_assert(kArmCodeAlignment == 8, "Code below assumes kArmCodeAlignment == 8");
constexpr size_t kBakerLiteralOffset1 = kArmCodeAlignment + 2u - kPcAdjustment;
constexpr size_t kBakerLiteralOffset2 = kBakerLiteralOffset1 + kArmCodeAlignment;
// Use offset = 0, base_reg = 0, the LDR is simply `kLdrWInsn | (root_reg << 12)`.
const uint32_t ldr1 = kLdrWInsn | (/* root_reg */ 1 << 12);
const uint32_t ldr2 = kLdrWInsn | (/* root_reg */ 2 << 12);
const std::vector<uint8_t> last_method_raw_code = RawCode({
kNopInsn, // Padding before first GC root read barrier.
ldr1, kBneWPlus0, // First GC root LDR with read barrier.
ldr2, kBneWPlus0, // Second GC root LDR with read barrier.
});
uint32_t encoded_data1 = EncodeBakerReadBarrierGcRootData(/* root_reg */ 1, /* narrow */ false);
uint32_t encoded_data2 = EncodeBakerReadBarrierGcRootData(/* root_reg */ 2, /* narrow */ false);
const LinkerPatch last_method_patches[] = {
LinkerPatch::BakerReadBarrierBranchPatch(kBakerLiteralOffset1, encoded_data1),
LinkerPatch::BakerReadBarrierBranchPatch(kBakerLiteralOffset2, encoded_data2),
};
++method_idx;
AddCompiledMethod(MethodRef(method_idx),
ArrayRef<const uint8_t>(last_method_raw_code),
ArrayRef<const LinkerPatch>(last_method_patches));
// The main purpose of the test is to check that Link() does not cause a crash.
Link();
ASSERT_EQ(15 * MB, GetMethodOffset(method_idx) - GetMethodOffset(1u));
}
} // namespace linker
} // namespace art