blob: be893efb6dbd94ef6b669d5e085fe3de13813d9b [file] [log] [blame]
/* Copyright (C) 2008 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* File: OP_SHR_LONG.S
*
* Code: Performs a shift right long
*
* For: shl-long
*
* Description: Perform a binary shift operation using two source registers
* where one is the shift amount and the other is the value to shift.
* Store the result in a destination register.
*
* Format: AA|op CC|BB (23x)
*
* Syntax: op vAA, vBB, vCC
*/
FETCH_BB 1, %edx # %edx<- BB
FETCH_CC 1, %eax # %eax<- CC
movq (rFP, %edx, 4), %xmm1 # %xmm1<- vBB
movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC
movq .LshiftMask, %xmm2
pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits
psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB
cmpl $$0, 4(rFP, %edx, 4) # check if we need to consider sign
jl .L${opcode}_finish # consider sign
jmp .L${opcode}_final # sign is fine, finish
%break
.L${opcode}_finish:
movq .Lvalue64, %xmm3 # %xmm3<- 64
psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
movq .L64bits, %xmm4 # %xmm4<- lower 64 bits set
psllq %xmm3, %xmm4 # %xmm4<- correct mask for sign bits
por %xmm4, %xmm1 # %xmm1<- signed and shifted vBB
.L${opcode}_final:
movq %xmm1, (rFP, rINST, 4) # vAA<- shifted vBB
FINISH 2 # jump to next instruction