blob: bfec8c022e0a59fceeba6b5459c6ed268d4d64f9 [file] [log] [blame]
/*
* Copyright (C) 2013 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_ARCH_X86_64_ASM_SUPPORT_X86_64_S_
#define ART_RUNTIME_ARCH_X86_64_ASM_SUPPORT_X86_64_S_
#include "asm_support_x86_64.h"
#include "interpreter/cfi_asm_support.h"
// Regular gas(1) & current clang/llvm assembler support named macro parameters.
#define MACRO0(macro_name) .macro macro_name
#define MACRO1(macro_name, macro_arg1) .macro macro_name macro_arg1
#define MACRO2(macro_name, macro_arg1, macro_arg2) .macro macro_name macro_arg1, macro_arg2
#define MACRO3(macro_name, macro_arg1, macro_arg2, macro_arg3) .macro macro_name macro_arg1, macro_arg2, macro_arg3
#define MACRO4(macro_name, macro_arg1, macro_arg2, macro_arg3, macro_arg4) .macro macro_name macro_arg1, macro_arg2, macro_arg3, macro_arg4
#define END_MACRO .endm
#if defined(__clang__)
// Clang/llvm does not support .altmacro. However, the clang/llvm preprocessor doesn't
// separate the backslash and parameter by a space. Everything just works.
#define RAW_VAR(name) \name
#define VAR(name) \name
#define CALLVAR(name) SYMBOL(\name)
#define PLT_VAR(name) \name@PLT
#define REG_VAR(name) %\name
#define CALL_MACRO(name) \name
#else
// Regular gas(1) uses \argument_name for macro arguments.
// We need to turn on alternate macro syntax so we can use & instead or the preprocessor
// will screw us by inserting a space between the \ and the name. Even in this mode there's
// no special meaning to $, so literals are still just $x. The use of altmacro means % is a
// special character meaning care needs to be taken when passing registers as macro
// arguments.
.altmacro
#define RAW_VAR(name) name&
#define VAR(name) name&
#define CALLVAR(name) SYMBOL(name&)
#define PLT_VAR(name) name&@PLT
#define REG_VAR(name) %name
#define CALL_MACRO(name) name&
#endif
#define LITERAL(value) $value
#if defined(__APPLE__)
#define MACRO_LITERAL(value) $(value)
#else
#define MACRO_LITERAL(value) $value
#endif
#if defined(__APPLE__)
#define FUNCTION_TYPE(name)
#define SIZE(name)
#else
#define FUNCTION_TYPE(name) .type name, @function
#define SIZE(name) .size name, .-name
#endif
// CFI support.
#if !defined(__APPLE__)
#define CFI_STARTPROC .cfi_startproc
#define CFI_ENDPROC .cfi_endproc
#define CFI_ADJUST_CFA_OFFSET(size) .cfi_adjust_cfa_offset size
#define CFI_DEF_CFA(reg,size) .cfi_def_cfa reg,size
#define CFI_DEF_CFA_REGISTER(reg) .cfi_def_cfa_register reg
#define CFI_RESTORE(reg) .cfi_restore reg
#define CFI_REL_OFFSET(reg,size) .cfi_rel_offset reg,size
#define CFI_REGISTER(orig_reg, current_reg) .cfi_register orig_reg, current_reg
#define CFI_REMEMBER_STATE .cfi_remember_state
// The spec is not clear whether the CFA is part of the saved state and tools
// differ in the behaviour, so explicitly set the CFA to avoid any ambiguity.
// The restored CFA state should match the CFA state during CFI_REMEMBER_STATE.
// `objdump -Wf libart.so | egrep "_cfa|_state"` is useful to audit the opcodes.
MACRO2(CFI_RESTORE_STATE_AND_DEF_CFA, reg, off)
.cfi_restore_state
.cfi_def_cfa \reg,\off
END_MACRO
#define CFI_RESTORE_STATE .cfi_restore_state
#else
// Mac OS' doesn't like cfi_* directives.
#define CFI_STARTPROC
#define CFI_ENDPROC
#define CFI_ADJUST_CFA_OFFSET(size)
#define CFI_DEF_CFA(reg,size)
#define CFI_DEF_CFA_REGISTER(reg)
#define CFI_RESTORE(reg)
#define CFI_REL_OFFSET(reg,size)
#define CFI_REGISTER(orig_reg, current_reg)
#define CFI_REMEMBER_STATE
#define CFI_RESTORE_STATE_AND_DEF_CFA(off)
MACRO2(CFI_RESTORE_STATE_AND_DEF_CFA, reg, off)
END_MACRO
#define CFI_RESTORE_STATE
#endif
// The register numbers are a bit mixed up for x86-64.
#define CFI_REG_rax 0
#define CFI_REG_rcx 2
#define CFI_REG_rdx 1
#define CFI_REG_rbx 3
#define CFI_REG_rsp 7
#define CFI_REG_rbp 6
#define CFI_REG_rsi 4
#define CFI_REG_rdi 5
#define CFI_REG_r8 8
#define CFI_REG_r9 9
#define CFI_REG_r10 10
#define CFI_REG_r11 11
#define CFI_REG_r12 12
#define CFI_REG_r13 13
#define CFI_REG_r14 14
#define CFI_REG_r15 15
#define CFI_REG_rip 16
#define CFI_REG(reg) CFI_REG_##reg
MACRO3(CFI_EXPRESSION_BREG, n, b, offset)
.if (-0x40 <= (\offset)) && ((\offset) < 0x40)
CFI_EXPRESSION_BREG_1(\n, \b, \offset)
.elseif (-0x2000 <= (\offset)) && ((\offset) < 0x2000)
CFI_EXPRESSION_BREG_2(\n, \b, \offset)
.else
.error "Unsupported offset"
.endif
END_MACRO
MACRO3(CFI_DEF_CFA_BREG_PLUS_UCONST, reg, offset, size)
.if ((\size) < 0)
.error "Size should be positive"
.endif
.if (((\offset) < -0x40) || ((\offset) >= 0x40))
.error "Unsupported offset"
.endif
.if ((\size) < 0x80)
CFI_DEF_CFA_BREG_PLUS_UCONST_1_1(\reg, \offset, \size)
.elseif ((\size) < 0x4000)
CFI_DEF_CFA_BREG_PLUS_UCONST_1_2(\reg, \offset, \size)
.else
.error "Unsupported size"
.endif
END_MACRO
// Symbols.
#if !defined(__APPLE__)
#define SYMBOL(name) name
#define PLT_SYMBOL(name) name ## @PLT
#else
#define SYMBOL(name) _ ## name
#define PLT_SYMBOL(name) _ ## name
#endif
// Directive to hide a function symbol.
#if defined(__APPLE__)
#define ASM_HIDDEN .private_extern
#else
#define ASM_HIDDEN .hidden
#endif
/* Cache alignment for function entry */
MACRO0(ALIGN_FUNCTION_ENTRY)
// ART-compiled functions have OatQuickMethodHeader but assembly funtions do not.
// Prefix the assembly code with 0xFFs, which means there is no method header.
.byte 0xFF, 0xFF, 0xFF, 0xFF
// Cache alignment for function entry.
.balign 16, 0xFF
END_MACRO
// TODO: we might need to use SYMBOL() here to add the underscore prefix
// for mac builds.
MACRO2(DEFINE_FUNCTION_CUSTOM_CFA, c_name, cfa_offset)
FUNCTION_TYPE(SYMBOL(\c_name))
ASM_HIDDEN CALLVAR(c_name)
.globl CALLVAR(c_name)
ALIGN_FUNCTION_ENTRY
CALLVAR(c_name):
CFI_STARTPROC
// Ensure we get an appropriate starting CFA.
CFI_DEF_CFA(rsp, RAW_VAR(cfa_offset))
END_MACRO
MACRO1(DEFINE_FUNCTION, c_name)
DEFINE_FUNCTION_CUSTOM_CFA RAW_VAR(c_name), __SIZEOF_POINTER__
END_MACRO
MACRO1(END_FUNCTION, c_name)
CFI_ENDPROC
SIZE(SYMBOL(\c_name))
END_MACRO
MACRO1(PUSH, reg)
pushq REG_VAR(reg)
CFI_ADJUST_CFA_OFFSET(8)
CFI_REL_OFFSET(REG_VAR(reg), 0)
END_MACRO
MACRO1(POP, reg)
popq REG_VAR(reg)
CFI_ADJUST_CFA_OFFSET(-8)
CFI_RESTORE(REG_VAR(reg))
END_MACRO
// Arguments do not need .cfi_rel_offset as they are caller-saved and
// therefore cannot hold caller's variables or unwinding data.
MACRO1(PUSH_ARG, reg)
pushq REG_VAR(reg)
CFI_ADJUST_CFA_OFFSET(8)
END_MACRO
MACRO1(POP_ARG, reg)
popq REG_VAR(reg)
CFI_ADJUST_CFA_OFFSET(-8)
END_MACRO
MACRO3(SAVE_REG_BASE, base, reg, offset)
movq REG_VAR(reg), RAW_VAR(offset)(REG_VAR(base))
CFI_REL_OFFSET(REG_VAR(reg), RAW_VAR(offset))
END_MACRO
MACRO3(RESTORE_REG_BASE, base, reg, offset)
movq RAW_VAR(offset)(REG_VAR(base)), REG_VAR(reg)
CFI_RESTORE(REG_VAR(reg))
END_MACRO
MACRO1(INCREASE_FRAME, frame_adjustment)
subq MACRO_LITERAL(RAW_VAR(frame_adjustment)), %rsp
CFI_ADJUST_CFA_OFFSET((RAW_VAR(frame_adjustment)))
END_MACRO
MACRO1(DECREASE_FRAME, frame_adjustment)
addq MACRO_LITERAL(RAW_VAR(frame_adjustment)), %rsp
CFI_ADJUST_CFA_OFFSET(-(RAW_VAR(frame_adjustment)))
END_MACRO
MACRO1(UNIMPLEMENTED,name)
FUNCTION_TYPE(SYMBOL(\name))
ASM_HIDDEN VAR(name)
.globl VAR(name)
ALIGN_FUNCTION_ENTRY
VAR(name):
CFI_STARTPROC
int3
int3
CFI_ENDPROC
SIZE(SYMBOL(\name))
END_MACRO
MACRO0(UNREACHABLE)
int3
END_MACRO
MACRO0(UNTESTED)
int3
END_MACRO
MACRO1(LOAD_RUNTIME_INSTANCE, reg)
movq _ZN3art7Runtime9instance_E@GOTPCREL(%rip), REG_VAR(reg)
movq (REG_VAR(reg)), REG_VAR(reg)
END_MACRO
// Macros to poison (negate) the reference for heap poisoning.
MACRO1(POISON_HEAP_REF, rRef)
#ifdef USE_HEAP_POISONING
negl REG_VAR(rRef)
#endif // USE_HEAP_POISONING
END_MACRO
// Macros to unpoison (negate) the reference for heap poisoning.
MACRO1(UNPOISON_HEAP_REF, rRef)
#ifdef USE_HEAP_POISONING
negl REG_VAR(rRef)
#endif // USE_HEAP_POISONING
END_MACRO
/*
* Macro that sets up the callee save frame to conform with
* Runtime::CreateCalleeSaveMethod(kSaveRefsOnly)
*/
MACRO0(SETUP_SAVE_REFS_ONLY_FRAME)
#if defined(__APPLE__)
int3
int3
#else
// R10 := Runtime::Current()
LOAD_RUNTIME_INSTANCE r10
// Save callee and GPR args, mixed together to agree with core spills bitmap.
PUSH r15 // Callee save.
PUSH r14 // Callee save.
PUSH r13 // Callee save.
PUSH r12 // Callee save.
PUSH rbp // Callee save.
PUSH rbx // Callee save.
// Create space for FPR args, plus space for ArtMethod*.
INCREASE_FRAME 8 + 4 * 8
// Save FPRs.
movq %xmm12, 8(%rsp)
movq %xmm13, 16(%rsp)
movq %xmm14, 24(%rsp)
movq %xmm15, 32(%rsp)
// R10 := ArtMethod* for refs only callee save frame method.
movq RUNTIME_SAVE_REFS_ONLY_METHOD_OFFSET(%r10), %r10
// Store ArtMethod* to bottom of stack.
movq %r10, 0(%rsp)
// Store rsp as the stop quick frame.
movq %rsp, %gs:THREAD_TOP_QUICK_FRAME_OFFSET
// Ugly compile-time check, but we only have the preprocessor.
// Last +8: implicit return address pushed on stack when caller made call.
#if (FRAME_SIZE_SAVE_REFS_ONLY != 6 * 8 + 4 * 8 + 8 + 8)
#error "FRAME_SIZE_SAVE_REFS_ONLY(X86_64) size not as expected."
#endif
#endif // __APPLE__
END_MACRO
MACRO0(RESTORE_SAVE_REFS_ONLY_FRAME)
movq 8(%rsp), %xmm12
movq 16(%rsp), %xmm13
movq 24(%rsp), %xmm14
movq 32(%rsp), %xmm15
DECREASE_FRAME 8 + 4*8
// TODO: optimize by not restoring callee-saves restored by the ABI
POP rbx
POP rbp
POP r12
POP r13
POP r14
POP r15
END_MACRO
/*
* Macro that sets up the callee save frame to conform with
* Runtime::CreateCalleeSaveMethod(kSaveRefsAndArgs), except for storing the method.
*/
MACRO0(SETUP_SAVE_REFS_AND_ARGS_FRAME_REGISTERS_ONLY)
// Save callee and GPR args, mixed together to agree with core spills bitmap.
PUSH r15 // Callee save.
PUSH r14 // Callee save.
PUSH r13 // Callee save.
PUSH r12 // Callee save.
PUSH_ARG r9 // Quick arg 5.
PUSH_ARG r8 // Quick arg 4.
PUSH_ARG rsi // Quick arg 1.
PUSH rbp // Callee save.
PUSH rbx // Callee save.
PUSH_ARG rdx // Quick arg 2.
PUSH_ARG rcx // Quick arg 3.
// Create space for FPR args and create 2 slots for ArtMethod*.
INCREASE_FRAME 16 + 12 * 8
// Save FPRs.
movq %xmm0, 16(%rsp)
movq %xmm1, 24(%rsp)
movq %xmm2, 32(%rsp)
movq %xmm3, 40(%rsp)
movq %xmm4, 48(%rsp)
movq %xmm5, 56(%rsp)
movq %xmm6, 64(%rsp)
movq %xmm7, 72(%rsp)
movq %xmm12, 80(%rsp)
movq %xmm13, 88(%rsp)
movq %xmm14, 96(%rsp)
movq %xmm15, 104(%rsp)
// Ugly compile-time check, but we only have the preprocessor.
// Last +8: implicit return address pushed on stack when caller made call.
#if (FRAME_SIZE_SAVE_REFS_AND_ARGS != 11 * 8 + 12 * 8 + 16 + 8)
#error "FRAME_SIZE_SAVE_REFS_AND_ARGS(X86_64) size not as expected."
#endif
END_MACRO
MACRO0(RESTORE_SAVE_REFS_AND_ARGS_FRAME)
// Restore FPRs.
movq 16(%rsp), %xmm0
movq 24(%rsp), %xmm1
movq 32(%rsp), %xmm2
movq 40(%rsp), %xmm3
movq 48(%rsp), %xmm4
movq 56(%rsp), %xmm5
movq 64(%rsp), %xmm6
movq 72(%rsp), %xmm7
movq 80(%rsp), %xmm12
movq 88(%rsp), %xmm13
movq 96(%rsp), %xmm14
movq 104(%rsp), %xmm15
DECREASE_FRAME 80 + 4 * 8
// Restore callee and GPR args, mixed together to agree with core spills bitmap.
POP_ARG rcx
POP_ARG rdx
POP rbx
POP rbp
POP_ARG rsi
POP_ARG r8
POP_ARG r9
POP r12
POP r13
POP r14
POP r15
END_MACRO
/*
* Macro that sets up the callee save frame to conform with
* Runtime::CreateCalleeSaveMethod(kSaveAllCalleeSaves)
*/
MACRO0(SETUP_SAVE_ALL_CALLEE_SAVES_FRAME)
#if defined(__APPLE__)
int3
int3
#else
// R10 := Runtime::Current()
LOAD_RUNTIME_INSTANCE r10
// Save callee save registers to agree with core spills bitmap.
PUSH r15 // Callee save.
PUSH r14 // Callee save.
PUSH r13 // Callee save.
PUSH r12 // Callee save.
PUSH rbp // Callee save.
PUSH rbx // Callee save.
// Create space for FPR args, plus space for ArtMethod*.
INCREASE_FRAME 4 * 8 + 8
// Save FPRs.
movq %xmm12, 8(%rsp)
movq %xmm13, 16(%rsp)
movq %xmm14, 24(%rsp)
movq %xmm15, 32(%rsp)
// R10 := ArtMethod* for save all callee save frame method.
movq RUNTIME_SAVE_ALL_CALLEE_SAVES_METHOD_OFFSET(%r10), %r10
// Store ArtMethod* to bottom of stack.
movq %r10, 0(%rsp)
// Store rsp as the top quick frame.
movq %rsp, %gs:THREAD_TOP_QUICK_FRAME_OFFSET
// Ugly compile-time check, but we only have the preprocessor.
// Last +8: implicit return address pushed on stack when caller made call.
#if (FRAME_SIZE_SAVE_ALL_CALLEE_SAVES != 6 * 8 + 4 * 8 + 8 + 8)
#error "FRAME_SIZE_SAVE_ALL_CALLEE_SAVES(X86_64) size not as expected."
#endif
#endif // __APPLE__
END_MACRO
MACRO0(SETUP_FP_CALLEE_SAVE_FRAME)
// Create space for ART FP callee-saved registers
INCREASE_FRAME 4 * 8
movq %xmm12, 0(%rsp)
movq %xmm13, 8(%rsp)
movq %xmm14, 16(%rsp)
movq %xmm15, 24(%rsp)
END_MACRO
MACRO0(RESTORE_FP_CALLEE_SAVE_FRAME)
// Restore ART FP callee-saved registers
movq 0(%rsp), %xmm12
movq 8(%rsp), %xmm13
movq 16(%rsp), %xmm14
movq 24(%rsp), %xmm15
DECREASE_FRAME 4 * 8
END_MACRO
/*
* Macro that calls through to artDeliverPendingExceptionFromCode, where the pending
* exception is Thread::Current()->exception_ when the runtime method frame is ready.
*/
MACRO0(DELIVER_PENDING_EXCEPTION_FRAME_READY)
// (Thread*) setup
movq %gs:THREAD_SELF_OFFSET, %rdi
call SYMBOL(artDeliverPendingExceptionFromCode) // artDeliverPendingExceptionFromCode(Thread*)
UNREACHABLE
END_MACRO
/*
* Macro that calls through to artDeliverPendingExceptionFromCode, where the pending
* exception is Thread::Current()->exception_.
*/
MACRO0(DELIVER_PENDING_EXCEPTION)
SETUP_SAVE_ALL_CALLEE_SAVES_FRAME // save callee saves for throw
DELIVER_PENDING_EXCEPTION_FRAME_READY
END_MACRO
MACRO0(RETURN_OR_DELIVER_PENDING_EXCEPTION)
movq %gs:THREAD_EXCEPTION_OFFSET, %rcx // get exception field
testq %rcx, %rcx // rcx == 0 ?
jnz 1f // if rcx != 0 goto 1
ret // return
1: // deliver exception on current thread
DELIVER_PENDING_EXCEPTION
END_MACRO
// Locking is needed for both managed code and JNI stubs.
MACRO3(LOCK_OBJECT_FAST_PATH, obj, tmp, slow_lock)
1:
movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word
movl %gs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp: thread id.
xorl %eax, REG_VAR(tmp) // tmp: thread id with count 0 + read barrier bits.
testl LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED), %eax // Test the non-gc bits.
jnz 2f // Check if unlocked.
// Unlocked case - store tmp: original lock word plus thread id, preserved read barrier bits.
lock cmpxchg REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
jnz 1b // cmpxchg failed retry
ret
2: // EAX: original lock word, tmp: thread id ^ EAX
// Check lock word state and thread id together,
testl LITERAL(LOCK_WORD_STATE_MASK_SHIFTED | LOCK_WORD_THIN_LOCK_OWNER_MASK_SHIFTED), \
REG_VAR(tmp)
jne \slow_lock // Slow path if either of the two high bits are set.
// Increment the recursive lock count.
leal LOCK_WORD_THIN_LOCK_COUNT_ONE(%eax), REG_VAR(tmp)
testl LITERAL(LOCK_WORD_THIN_LOCK_COUNT_MASK_SHIFTED), REG_VAR(tmp)
je \slow_lock // If count overflowed, go to slow lock.
// Update lockword for recursive lock, cmpxchg necessary for read barrier bits.
// EAX: old val, tmp: new val.
lock cmpxchg REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
jnz 1b // cmpxchg failed retry
ret
END_MACRO
// Unlocking is needed for both managed code and JNI stubs.
MACRO4(UNLOCK_OBJECT_FAST_PATH, obj, tmp, saved_rax, slow_unlock)
1:
movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word
movl %gs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp := thread id
xorl %eax, REG_VAR(tmp) // tmp := thread id ^ lock word
test LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED), REG_VAR(tmp)
jnz 2f // Check if simply locked.
// Transition to unlocked.
#ifndef USE_READ_BARRIER
movl REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
#else
lock cmpxchg REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
jnz 1b // cmpxchg failed retry
#endif
.ifnc \saved_rax, none
movq REG_VAR(saved_rax), %rax // Restore RAX.
.endif
ret
2: // EAX: original lock word, tmp: lock_word ^ thread id
// Check lock word state and thread id together.
testl LITERAL(LOCK_WORD_STATE_MASK_SHIFTED | LOCK_WORD_THIN_LOCK_OWNER_MASK_SHIFTED), \
REG_VAR(tmp)
jnz \slow_unlock
// Update lockword for recursive unlock, cmpxchg necessary for read barrier bits.
// tmp: new lock word with decremented count.
leal -LOCK_WORD_THIN_LOCK_COUNT_ONE(%eax), REG_VAR(tmp)
#ifndef USE_READ_BARRIER
// EAX: new lock word with decremented count.
movl REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
#else
lock cmpxchg REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
jnz 1b // cmpxchg failed retry
#endif
.ifnc \saved_rax, none
movq REG_VAR(saved_rax), %rax // Restore RAX.
.endif
ret
END_MACRO
#endif // ART_RUNTIME_ARCH_X86_64_ASM_SUPPORT_X86_64_S_