blob: e4caa579a526f86c2949279ec32a2156771fc052 [file] [log] [blame]
/*--------------------------------------------------------------------*/
/*--- begin guest_ppc_toIR.c ---*/
/*--------------------------------------------------------------------*/
/*
This file is part of Valgrind, a dynamic binary instrumentation
framework.
Copyright (C) 2004-2011 OpenWorks LLP
info@open-works.net
This program is free software; you can redistribute it and/or
modify it under the terms of the GNU General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
02110-1301, USA.
The GNU General Public License is contained in the file COPYING.
Neither the names of the U.S. Department of Energy nor the
University of California nor the names of its contributors may be
used to endorse or promote products derived from this software
without prior written permission.
*/
/* TODO 18/Nov/05:
Spot rld... cases which are simply left/right shifts and emit
Shl64/Shr64 accordingly.
Altivec
- datastream insns
- lvxl,stvxl: load/store with 'least recently used' hint
- vexptefp, vlogefp
LIMITATIONS:
Various, including:
- Some invalid forms of lswi and lswx are accepted when they should
not be.
- Floating Point:
- All exceptions disabled in FPSCR
- condition codes not set in FPSCR
- Altivec floating point:
- vmaddfp, vnmsubfp
Because we're using Java/IEEE mode (FPSCR[NJ]), rather than the
system default of Non-Java mode, we get some small errors
(lowest bit only).
This is because Non-Java mode brutally hacks denormalised results
to zero, whereas we keep maximum accuracy. However, using
Non-Java mode would give us more inaccuracy, as our intermediate
results would then be zeroed, too.
- AbiHints for the stack red zone are only emitted for
unconditional calls and returns (bl, blr). They should also be
emitted for conditional calls and returns, but we don't have a
way to express that right now. Ah well.
*/
/* "Special" instructions.
This instruction decoder can decode four special instructions
which mean nothing natively (are no-ops as far as regs/mem are
concerned) but have meaning for supporting Valgrind. A special
instruction is flagged by a 16-byte preamble:
32-bit mode: 54001800 54006800 5400E800 54009800
(rlwinm 0,0,3,0,0; rlwinm 0,0,13,0,0;
rlwinm 0,0,29,0,0; rlwinm 0,0,19,0,0)
64-bit mode: 78001800 78006800 7800E802 78009802
(rotldi 0,0,3; rotldi 0,0,13;
rotldi 0,0,61; rotldi 0,0,51)
Following that, one of the following 3 are allowed
(standard interpretation in parentheses):
7C210B78 (or 1,1,1) %R3 = client_request ( %R4 )
7C421378 (or 2,2,2) %R3 = guest_NRADDR
7C631B78 (or 3,3,3) branch-and-link-to-noredir %R11
7C842378 (or 4,4,4) %R3 = guest_NRADDR_GPR2
Any other bytes following the 16-byte preamble are illegal and
constitute a failure in instruction decoding. This all assumes
that the preamble will never occur except in specific code
fragments designed for Valgrind to catch.
*/
/* Translates PPC32/64 code to IR. */
/* References
#define PPC32
"PowerPC Microprocessor Family:
The Programming Environments Manual for 32-Bit Microprocessors"
02/21/2000
http://www-3.ibm.com/chips/techlib/techlib.nsf/techdocs/852569B20050FF778525699600719DF2
#define PPC64
"PowerPC Microprocessor Family:
Programming Environments Manual for 64-Bit Microprocessors"
06/10/2003
http://www-3.ibm.com/chips/techlib/techlib.nsf/techdocs/F7E732FF811F783187256FDD004D3797
#define AV
"PowerPC Microprocessor Family:
AltiVec(TM) Technology Programming Environments Manual"
07/10/2003
http://www-3.ibm.com/chips/techlib/techlib.nsf/techdocs/FBFA164F824370F987256D6A006F424D
*/
#include "libvex_basictypes.h"
#include "libvex_ir.h"
#include "libvex.h"
#include "libvex_guest_ppc32.h"
#include "libvex_guest_ppc64.h"
#include "main_util.h"
#include "main_globals.h"
#include "guest_generic_bb_to_IR.h"
#include "guest_ppc_defs.h"
/*------------------------------------------------------------*/
/*--- Globals ---*/
/*------------------------------------------------------------*/
/* These are set at the start of the translation of an insn, right
down in disInstr_PPC, so that we don't have to pass them around
endlessly. They are all constant during the translation of any
given insn. */
/* We need to know this to do sub-register accesses correctly. */
static Bool host_is_bigendian;
/* Pointer to the guest code area. */
static UChar* guest_code;
/* The guest address corresponding to guest_code[0]. */
static Addr64 guest_CIA_bbstart;
/* The guest address for the instruction currently being
translated. */
static Addr64 guest_CIA_curr_instr;
/* The IRSB* into which we're generating code. */
static IRSB* irsb;
/* Is our guest binary 32 or 64bit? Set at each call to
disInstr_PPC below. */
static Bool mode64 = False;
// Given a pointer to a function as obtained by "& functionname" in C,
// produce a pointer to the actual entry point for the function. For
// most platforms it's the identity function. Unfortunately, on
// ppc64-linux it isn't (sigh) and ditto for ppc32-aix5 and
// ppc64-aix5.
static void* fnptr_to_fnentry( VexAbiInfo* vbi, void* f )
{
if (vbi->host_ppc_calls_use_fndescrs) {
/* f is a pointer to a 3-word function descriptor, of which the
first word is the entry address. */
/* note, this is correct even with cross-jitting, since this is
purely a host issue, not a guest one. */
HWord* fdescr = (HWord*)f;
return (void*)(fdescr[0]);
} else {
/* Simple; "& f" points directly at the code for f. */
return f;
}
}
#define SIGN_BIT 0x8000000000000000ULL
#define SIGN_MASK 0x7fffffffffffffffULL
#define SIGN_BIT32 0x80000000
#define SIGN_MASK32 0x7fffffff
/*------------------------------------------------------------*/
/*--- Debugging output ---*/
/*------------------------------------------------------------*/
#define DIP(format, args...) \
if (vex_traceflags & VEX_TRACE_FE) \
vex_printf(format, ## args)
#define DIS(buf, format, args...) \
if (vex_traceflags & VEX_TRACE_FE) \
vex_sprintf(buf, format, ## args)
/*------------------------------------------------------------*/
/*--- Offsets of various parts of the ppc32/64 guest state ---*/
/*------------------------------------------------------------*/
#define offsetofPPCGuestState(_x) \
(mode64 ? offsetof(VexGuestPPC64State, _x) : \
offsetof(VexGuestPPC32State, _x))
#define OFFB_CIA offsetofPPCGuestState(guest_CIA)
#define OFFB_IP_AT_SYSCALL offsetofPPCGuestState(guest_IP_AT_SYSCALL)
#define OFFB_SPRG3_RO offsetofPPCGuestState(guest_SPRG3_RO)
#define OFFB_LR offsetofPPCGuestState(guest_LR)
#define OFFB_CTR offsetofPPCGuestState(guest_CTR)
#define OFFB_XER_SO offsetofPPCGuestState(guest_XER_SO)
#define OFFB_XER_OV offsetofPPCGuestState(guest_XER_OV)
#define OFFB_XER_CA offsetofPPCGuestState(guest_XER_CA)
#define OFFB_XER_BC offsetofPPCGuestState(guest_XER_BC)
#define OFFB_FPROUND offsetofPPCGuestState(guest_FPROUND)
#define OFFB_DFPROUND offsetofPPCGuestState(guest_DFPROUND)
#define OFFB_VRSAVE offsetofPPCGuestState(guest_VRSAVE)
#define OFFB_VSCR offsetofPPCGuestState(guest_VSCR)
#define OFFB_EMWARN offsetofPPCGuestState(guest_EMWARN)
#define OFFB_TISTART offsetofPPCGuestState(guest_TISTART)
#define OFFB_TILEN offsetofPPCGuestState(guest_TILEN)
#define OFFB_NRADDR offsetofPPCGuestState(guest_NRADDR)
#define OFFB_NRADDR_GPR2 offsetofPPCGuestState(guest_NRADDR_GPR2)
/*------------------------------------------------------------*/
/*--- Extract instruction fields --- */
/*------------------------------------------------------------*/
/* Extract field from insn, given idx (zero = lsb) and field length */
#define IFIELD( insn, idx, len ) ((insn >> idx) & ((1<<len)-1))
/* Extract primary opcode, instr[31:26] */
static UChar ifieldOPC( UInt instr ) {
return toUChar( IFIELD( instr, 26, 6 ) );
}
/* Extract 10-bit secondary opcode, instr[10:1] */
static UInt ifieldOPClo10 ( UInt instr) {
return IFIELD( instr, 1, 10 );
}
/* Extract 9-bit secondary opcode, instr[9:1] */
static UInt ifieldOPClo9 ( UInt instr) {
return IFIELD( instr, 1, 9 );
}
/* Extract 8-bit secondary opcode, instr[8:1] */
static UInt ifieldOPClo8 ( UInt instr) {
return IFIELD( instr, 1, 8 );
}
/* Extract 5-bit secondary opcode, instr[5:1] */
static UInt ifieldOPClo5 ( UInt instr) {
return IFIELD( instr, 1, 5 );
}
/* Extract RD (destination register) field, instr[25:21] */
static UChar ifieldRegDS( UInt instr ) {
return toUChar( IFIELD( instr, 21, 5 ) );
}
/* Extract XT (destination register) field, instr[0,25:21] */
static UChar ifieldRegXT ( UInt instr )
{
UChar upper_bit = toUChar (IFIELD (instr, 0, 1));
UChar lower_bits = toUChar (IFIELD (instr, 21, 5));
return (upper_bit << 5) | lower_bits;
}
/* Extract XS (store source register) field, instr[0,25:21] */
static inline UChar ifieldRegXS ( UInt instr )
{
return ifieldRegXT ( instr );
}
/* Extract RA (1st source register) field, instr[20:16] */
static UChar ifieldRegA ( UInt instr ) {
return toUChar( IFIELD( instr, 16, 5 ) );
}
/* Extract XA (1st source register) field, instr[2,20:16] */
static UChar ifieldRegXA ( UInt instr )
{
UChar upper_bit = toUChar (IFIELD (instr, 2, 1));
UChar lower_bits = toUChar (IFIELD (instr, 16, 5));
return (upper_bit << 5) | lower_bits;
}
/* Extract RB (2nd source register) field, instr[15:11] */
static UChar ifieldRegB ( UInt instr ) {
return toUChar( IFIELD( instr, 11, 5 ) );
}
/* Extract XB (2nd source register) field, instr[1,15:11] */
static UChar ifieldRegXB ( UInt instr )
{
UChar upper_bit = toUChar (IFIELD (instr, 1, 1));
UChar lower_bits = toUChar (IFIELD (instr, 11, 5));
return (upper_bit << 5) | lower_bits;
}
/* Extract RC (3rd source register) field, instr[10:6] */
static UChar ifieldRegC ( UInt instr ) {
return toUChar( IFIELD( instr, 6, 5 ) );
}
/* Extract XC (3rd source register) field, instr[3,10:6] */
static UChar ifieldRegXC ( UInt instr )
{
UChar upper_bit = toUChar (IFIELD (instr, 3, 1));
UChar lower_bits = toUChar (IFIELD (instr, 6, 5));
return (upper_bit << 5) | lower_bits;
}
/* Extract bit 10, instr[10] */
static UChar ifieldBIT10 ( UInt instr ) {
return toUChar( IFIELD( instr, 10, 1 ) );
}
/* Extract 2nd lowest bit, instr[1] */
static UChar ifieldBIT1 ( UInt instr ) {
return toUChar( IFIELD( instr, 1, 1 ) );
}
/* Extract lowest bit, instr[0] */
static UChar ifieldBIT0 ( UInt instr ) {
return toUChar( instr & 0x1 );
}
/* Extract unsigned bottom half, instr[15:0] */
static UInt ifieldUIMM16 ( UInt instr ) {
return instr & 0xFFFF;
}
/* Extract unsigned bottom 26 bits, instr[25:0] */
static UInt ifieldUIMM26 ( UInt instr ) {
return instr & 0x3FFFFFF;
}
/* Extract DM field, instr[9:8] */
static UChar ifieldDM ( UInt instr ) {
return toUChar( IFIELD( instr, 8, 2 ) );
}
/* Extract SHW field, instr[9:8] */
static inline UChar ifieldSHW ( UInt instr )
{
return ifieldDM ( instr );
}
/*------------------------------------------------------------*/
/*--- Guest-state identifiers ---*/
/*------------------------------------------------------------*/
typedef enum {
PPC_GST_CIA, // Current Instruction Address
PPC_GST_LR, // Link Register
PPC_GST_CTR, // Count Register
PPC_GST_XER, // Overflow, carry flags, byte count
PPC_GST_CR, // Condition Register
PPC_GST_FPSCR, // Floating Point Status/Control Register
PPC_GST_VRSAVE, // Vector Save/Restore Register
PPC_GST_VSCR, // Vector Status and Control Register
PPC_GST_EMWARN, // Emulation warnings
PPC_GST_TISTART,// For icbi: start of area to invalidate
PPC_GST_TILEN, // For icbi: length of area to invalidate
PPC_GST_IP_AT_SYSCALL, // the CIA of the most recently executed SC insn
PPC_GST_SPRG3_RO, // SPRG3
PPC_GST_MAX
} PPC_GST;
#define MASK_FPSCR_RN 0x3ULL // Binary floating point rounding mode
#define MASK_FPSCR_DRN 0x700000000ULL // Decimal floating point rounding mode
#define MASK_VSCR_VALID 0x00010001
/*------------------------------------------------------------*/
/*--- FP Helpers ---*/
/*------------------------------------------------------------*/
/* Produce the 32-bit pattern corresponding to the supplied
float. */
static UInt float_to_bits ( Float f )
{
union { UInt i; Float f; } u;
vassert(4 == sizeof(UInt));
vassert(4 == sizeof(Float));
vassert(4 == sizeof(u));
u.f = f;
return u.i;
}
/*------------------------------------------------------------*/
/*--- Misc Helpers ---*/
/*------------------------------------------------------------*/
/* Generate mask with 1's from 'begin' through 'end',
wrapping if begin > end.
begin->end works from right to left, 0=lsb
*/
static UInt MASK32( UInt begin, UInt end )
{
UInt m1, m2, mask;
vassert(begin < 32);
vassert(end < 32);
m1 = ((UInt)(-1)) << begin;
m2 = ((UInt)(-1)) << end << 1;
mask = m1 ^ m2;
if (begin > end) mask = ~mask; // wrap mask
return mask;
}
static ULong MASK64( UInt begin, UInt end )
{
ULong m1, m2, mask;
vassert(begin < 64);
vassert(end < 64);
m1 = ((ULong)(-1)) << begin;
m2 = ((ULong)(-1)) << end << 1;
mask = m1 ^ m2;
if (begin > end) mask = ~mask; // wrap mask
return mask;
}
static Addr64 nextInsnAddr( void )
{
return guest_CIA_curr_instr + 4;
}
/*------------------------------------------------------------*/
/*--- Helper bits and pieces for deconstructing the ---*/
/*--- ppc32/64 insn stream. ---*/
/*------------------------------------------------------------*/
/* Add a statement to the list held by "irsb". */
static void stmt ( IRStmt* st )
{
addStmtToIRSB( irsb, st );
}
/* Generate a new temporary of the given type. */
static IRTemp newTemp ( IRType ty )
{
vassert(isPlausibleIRType(ty));
return newIRTemp( irsb->tyenv, ty );
}
/* Various simple conversions */
static UChar extend_s_5to8 ( UChar x )
{
return toUChar((((Int)x) << 27) >> 27);
}
static UInt extend_s_8to32( UChar x )
{
return (UInt)((((Int)x) << 24) >> 24);
}
static UInt extend_s_16to32 ( UInt x )
{
return (UInt)((((Int)x) << 16) >> 16);
}
static ULong extend_s_16to64 ( UInt x )
{
return (ULong)((((Long)x) << 48) >> 48);
}
static ULong extend_s_26to64 ( UInt x )
{
return (ULong)((((Long)x) << 38) >> 38);
}
static ULong extend_s_32to64 ( UInt x )
{
return (ULong)((((Long)x) << 32) >> 32);
}
/* Do a big-endian load of a 32-bit word, regardless of the endianness
of the underlying host. */
static UInt getUIntBigendianly ( UChar* p )
{
UInt w = 0;
w = (w << 8) | p[0];
w = (w << 8) | p[1];
w = (w << 8) | p[2];
w = (w << 8) | p[3];
return w;
}
/*------------------------------------------------------------*/
/*--- Helpers for constructing IR. ---*/
/*------------------------------------------------------------*/
static void assign ( IRTemp dst, IRExpr* e )
{
stmt( IRStmt_WrTmp(dst, e) );
}
/* This generates a normal (non store-conditional) store. */
static void storeBE ( IRExpr* addr, IRExpr* data )
{
IRType tyA = typeOfIRExpr(irsb->tyenv, addr);
vassert(tyA == Ity_I32 || tyA == Ity_I64);
stmt( IRStmt_Store(Iend_BE, addr, data) );
}
static IRExpr* unop ( IROp op, IRExpr* a )
{
return IRExpr_Unop(op, a);
}
static IRExpr* binop ( IROp op, IRExpr* a1, IRExpr* a2 )
{
return IRExpr_Binop(op, a1, a2);
}
static IRExpr* triop ( IROp op, IRExpr* a1, IRExpr* a2, IRExpr* a3 )
{
return IRExpr_Triop(op, a1, a2, a3);
}
static IRExpr* qop ( IROp op, IRExpr* a1, IRExpr* a2,
IRExpr* a3, IRExpr* a4 )
{
return IRExpr_Qop(op, a1, a2, a3, a4);
}
static IRExpr* mkexpr ( IRTemp tmp )
{
return IRExpr_RdTmp(tmp);
}
static IRExpr* mkU8 ( UChar i )
{
return IRExpr_Const(IRConst_U8(i));
}
static IRExpr* mkU16 ( UInt i )
{
return IRExpr_Const(IRConst_U16(i));
}
static IRExpr* mkU32 ( UInt i )
{
return IRExpr_Const(IRConst_U32(i));
}
static IRExpr* mkU64 ( ULong i )
{
return IRExpr_Const(IRConst_U64(i));
}
static IRExpr* mkV128 ( UShort i )
{
vassert(i == 0 || i == 0xffff);
return IRExpr_Const(IRConst_V128(i));
}
/* This generates a normal (non load-linked) load. */
static IRExpr* loadBE ( IRType ty, IRExpr* addr )
{
return IRExpr_Load(Iend_BE, ty, addr);
}
static IRExpr* mkOR1 ( IRExpr* arg1, IRExpr* arg2 )
{
vassert(typeOfIRExpr(irsb->tyenv, arg1) == Ity_I1);
vassert(typeOfIRExpr(irsb->tyenv, arg2) == Ity_I1);
return unop(Iop_32to1, binop(Iop_Or32, unop(Iop_1Uto32, arg1),
unop(Iop_1Uto32, arg2)));
}
static IRExpr* mkAND1 ( IRExpr* arg1, IRExpr* arg2 )
{
vassert(typeOfIRExpr(irsb->tyenv, arg1) == Ity_I1);
vassert(typeOfIRExpr(irsb->tyenv, arg2) == Ity_I1);
return unop(Iop_32to1, binop(Iop_And32, unop(Iop_1Uto32, arg1),
unop(Iop_1Uto32, arg2)));
}
/* expand V128_8Ux16 to 2x V128_16Ux8's */
static void expand8Ux16( IRExpr* vIn,
/*OUTs*/ IRTemp* vEvn, IRTemp* vOdd )
{
IRTemp ones8x16 = newTemp(Ity_V128);
vassert(typeOfIRExpr(irsb->tyenv, vIn) == Ity_V128);
vassert(vEvn && *vEvn == IRTemp_INVALID);
vassert(vOdd && *vOdd == IRTemp_INVALID);
*vEvn = newTemp(Ity_V128);
*vOdd = newTemp(Ity_V128);
assign( ones8x16, unop(Iop_Dup8x16, mkU8(0x1)) );
assign( *vOdd, binop(Iop_MullEven8Ux16, mkexpr(ones8x16), vIn) );
assign( *vEvn, binop(Iop_MullEven8Ux16, mkexpr(ones8x16),
binop(Iop_ShrV128, vIn, mkU8(8))) );
}
/* expand V128_8Sx16 to 2x V128_16Sx8's */
static void expand8Sx16( IRExpr* vIn,
/*OUTs*/ IRTemp* vEvn, IRTemp* vOdd )
{
IRTemp ones8x16 = newTemp(Ity_V128);
vassert(typeOfIRExpr(irsb->tyenv, vIn) == Ity_V128);
vassert(vEvn && *vEvn == IRTemp_INVALID);
vassert(vOdd && *vOdd == IRTemp_INVALID);
*vEvn = newTemp(Ity_V128);
*vOdd = newTemp(Ity_V128);
assign( ones8x16, unop(Iop_Dup8x16, mkU8(0x1)) );
assign( *vOdd, binop(Iop_MullEven8Sx16, mkexpr(ones8x16), vIn) );
assign( *vEvn, binop(Iop_MullEven8Sx16, mkexpr(ones8x16),
binop(Iop_ShrV128, vIn, mkU8(8))) );
}
/* expand V128_16Uto8 to 2x V128_32Ux4's */
static void expand16Ux8( IRExpr* vIn,
/*OUTs*/ IRTemp* vEvn, IRTemp* vOdd )
{
IRTemp ones16x8 = newTemp(Ity_V128);
vassert(typeOfIRExpr(irsb->tyenv, vIn) == Ity_V128);
vassert(vEvn && *vEvn == IRTemp_INVALID);
vassert(vOdd && *vOdd == IRTemp_INVALID);
*vEvn = newTemp(Ity_V128);
*vOdd = newTemp(Ity_V128);
assign( ones16x8, unop(Iop_Dup16x8, mkU16(0x1)) );
assign( *vOdd, binop(Iop_MullEven16Ux8, mkexpr(ones16x8), vIn) );
assign( *vEvn, binop(Iop_MullEven16Ux8, mkexpr(ones16x8),
binop(Iop_ShrV128, vIn, mkU8(16))) );
}
/* expand V128_16Sto8 to 2x V128_32Sx4's */
static void expand16Sx8( IRExpr* vIn,
/*OUTs*/ IRTemp* vEvn, IRTemp* vOdd )
{
IRTemp ones16x8 = newTemp(Ity_V128);
vassert(typeOfIRExpr(irsb->tyenv, vIn) == Ity_V128);
vassert(vEvn && *vEvn == IRTemp_INVALID);
vassert(vOdd && *vOdd == IRTemp_INVALID);
*vEvn = newTemp(Ity_V128);
*vOdd = newTemp(Ity_V128);
assign( ones16x8, unop(Iop_Dup16x8, mkU16(0x1)) );
assign( *vOdd, binop(Iop_MullEven16Sx8, mkexpr(ones16x8), vIn) );
assign( *vEvn, binop(Iop_MullEven16Sx8, mkexpr(ones16x8),
binop(Iop_ShrV128, vIn, mkU8(16))) );
}
/* break V128 to 4xF64's*/
static void breakV128to4xF64( IRExpr* t128,
/*OUTs*/
IRTemp* t3, IRTemp* t2,
IRTemp* t1, IRTemp* t0 )
{
IRTemp hi64 = newTemp(Ity_I64);
IRTemp lo64 = newTemp(Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t128) == Ity_V128);
vassert(t0 && *t0 == IRTemp_INVALID);
vassert(t1 && *t1 == IRTemp_INVALID);
vassert(t2 && *t2 == IRTemp_INVALID);
vassert(t3 && *t3 == IRTemp_INVALID);
*t0 = newTemp(Ity_F64);
*t1 = newTemp(Ity_F64);
*t2 = newTemp(Ity_F64);
*t3 = newTemp(Ity_F64);
assign( hi64, unop(Iop_V128HIto64, t128) );
assign( lo64, unop(Iop_V128to64, t128) );
assign( *t3,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32,
unop( Iop_64HIto32, mkexpr( hi64 ) ) ) ) );
assign( *t2,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, unop( Iop_64to32, mkexpr( hi64 ) ) ) ) );
assign( *t1,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32,
unop( Iop_64HIto32, mkexpr( lo64 ) ) ) ) );
assign( *t0,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, unop( Iop_64to32, mkexpr( lo64 ) ) ) ) );
}
/* break V128 to 4xI32's, then sign-extend to I64's */
static void breakV128to4x64S( IRExpr* t128,
/*OUTs*/
IRTemp* t3, IRTemp* t2,
IRTemp* t1, IRTemp* t0 )
{
IRTemp hi64 = newTemp(Ity_I64);
IRTemp lo64 = newTemp(Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t128) == Ity_V128);
vassert(t0 && *t0 == IRTemp_INVALID);
vassert(t1 && *t1 == IRTemp_INVALID);
vassert(t2 && *t2 == IRTemp_INVALID);
vassert(t3 && *t3 == IRTemp_INVALID);
*t0 = newTemp(Ity_I64);
*t1 = newTemp(Ity_I64);
*t2 = newTemp(Ity_I64);
*t3 = newTemp(Ity_I64);
assign( hi64, unop(Iop_V128HIto64, t128) );
assign( lo64, unop(Iop_V128to64, t128) );
assign( *t3, unop(Iop_32Sto64, unop(Iop_64HIto32, mkexpr(hi64))) );
assign( *t2, unop(Iop_32Sto64, unop(Iop_64to32, mkexpr(hi64))) );
assign( *t1, unop(Iop_32Sto64, unop(Iop_64HIto32, mkexpr(lo64))) );
assign( *t0, unop(Iop_32Sto64, unop(Iop_64to32, mkexpr(lo64))) );
}
/* break V128 to 4xI32's, then zero-extend to I64's */
static void breakV128to4x64U ( IRExpr* t128,
/*OUTs*/
IRTemp* t3, IRTemp* t2,
IRTemp* t1, IRTemp* t0 )
{
IRTemp hi64 = newTemp(Ity_I64);
IRTemp lo64 = newTemp(Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t128) == Ity_V128);
vassert(t0 && *t0 == IRTemp_INVALID);
vassert(t1 && *t1 == IRTemp_INVALID);
vassert(t2 && *t2 == IRTemp_INVALID);
vassert(t3 && *t3 == IRTemp_INVALID);
*t0 = newTemp(Ity_I64);
*t1 = newTemp(Ity_I64);
*t2 = newTemp(Ity_I64);
*t3 = newTemp(Ity_I64);
assign( hi64, unop(Iop_V128HIto64, t128) );
assign( lo64, unop(Iop_V128to64, t128) );
assign( *t3, unop(Iop_32Uto64, unop(Iop_64HIto32, mkexpr(hi64))) );
assign( *t2, unop(Iop_32Uto64, unop(Iop_64to32, mkexpr(hi64))) );
assign( *t1, unop(Iop_32Uto64, unop(Iop_64HIto32, mkexpr(lo64))) );
assign( *t0, unop(Iop_32Uto64, unop(Iop_64to32, mkexpr(lo64))) );
}
static void breakV128to4x32( IRExpr* t128,
/*OUTs*/
IRTemp* t3, IRTemp* t2,
IRTemp* t1, IRTemp* t0 )
{
IRTemp hi64 = newTemp(Ity_I64);
IRTemp lo64 = newTemp(Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t128) == Ity_V128);
vassert(t0 && *t0 == IRTemp_INVALID);
vassert(t1 && *t1 == IRTemp_INVALID);
vassert(t2 && *t2 == IRTemp_INVALID);
vassert(t3 && *t3 == IRTemp_INVALID);
*t0 = newTemp(Ity_I32);
*t1 = newTemp(Ity_I32);
*t2 = newTemp(Ity_I32);
*t3 = newTemp(Ity_I32);
assign( hi64, unop(Iop_V128HIto64, t128) );
assign( lo64, unop(Iop_V128to64, t128) );
assign( *t3, unop(Iop_64HIto32, mkexpr(hi64)) );
assign( *t2, unop(Iop_64to32, mkexpr(hi64)) );
assign( *t1, unop(Iop_64HIto32, mkexpr(lo64)) );
assign( *t0, unop(Iop_64to32, mkexpr(lo64)) );
}
/* Signed saturating narrow 64S to 32 */
static IRExpr* mkQNarrow64Sto32 ( IRExpr* t64 )
{
IRTemp hi32 = newTemp(Ity_I32);
IRTemp lo32 = newTemp(Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv, t64) == Ity_I64);
assign( hi32, unop(Iop_64HIto32, t64));
assign( lo32, unop(Iop_64to32, t64));
return IRExpr_Mux0X(
/* if (hi32 == (lo32 >>s 31)) */
unop(Iop_1Uto8,
binop(Iop_CmpEQ32, mkexpr(hi32),
binop( Iop_Sar32, mkexpr(lo32), mkU8(31)))),
/* else: sign dep saturate: 1->0x80000000, 0->0x7FFFFFFF */
binop(Iop_Add32, mkU32(0x7FFFFFFF),
binop(Iop_Shr32, mkexpr(hi32), mkU8(31))),
/* then: within signed-32 range: lo half good enough */
mkexpr(lo32) );
}
/* Unsigned saturating narrow 64S to 32 */
static IRExpr* mkQNarrow64Uto32 ( IRExpr* t64 )
{
IRTemp hi32 = newTemp(Ity_I32);
IRTemp lo32 = newTemp(Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv, t64) == Ity_I64);
assign( hi32, unop(Iop_64HIto32, t64));
assign( lo32, unop(Iop_64to32, t64));
return IRExpr_Mux0X(
/* if (top 32 bits of t64 are 0) */
unop(Iop_1Uto8, binop(Iop_CmpEQ32, mkexpr(hi32), mkU32(0))),
/* else: positive saturate -> 0xFFFFFFFF */
mkU32(0xFFFFFFFF),
/* then: within unsigned-32 range: lo half good enough */
mkexpr(lo32) );
}
/* Signed saturate narrow 64->32, combining to V128 */
static IRExpr* mkV128from4x64S ( IRExpr* t3, IRExpr* t2,
IRExpr* t1, IRExpr* t0 )
{
vassert(typeOfIRExpr(irsb->tyenv, t3) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t2) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t1) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t0) == Ity_I64);
return binop(Iop_64HLtoV128,
binop(Iop_32HLto64,
mkQNarrow64Sto32( t3 ),
mkQNarrow64Sto32( t2 )),
binop(Iop_32HLto64,
mkQNarrow64Sto32( t1 ),
mkQNarrow64Sto32( t0 )));
}
/* Unsigned saturate narrow 64->32, combining to V128 */
static IRExpr* mkV128from4x64U ( IRExpr* t3, IRExpr* t2,
IRExpr* t1, IRExpr* t0 )
{
vassert(typeOfIRExpr(irsb->tyenv, t3) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t2) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t1) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv, t0) == Ity_I64);
return binop(Iop_64HLtoV128,
binop(Iop_32HLto64,
mkQNarrow64Uto32( t3 ),
mkQNarrow64Uto32( t2 )),
binop(Iop_32HLto64,
mkQNarrow64Uto32( t1 ),
mkQNarrow64Uto32( t0 )));
}
/* Simulate irops Iop_MullOdd*, since we don't have them */
#define MK_Iop_MullOdd8Ux16( expr_vA, expr_vB ) \
binop(Iop_MullEven8Ux16, \
binop(Iop_ShrV128, expr_vA, mkU8(8)), \
binop(Iop_ShrV128, expr_vB, mkU8(8)))
#define MK_Iop_MullOdd8Sx16( expr_vA, expr_vB ) \
binop(Iop_MullEven8Sx16, \
binop(Iop_ShrV128, expr_vA, mkU8(8)), \
binop(Iop_ShrV128, expr_vB, mkU8(8)))
#define MK_Iop_MullOdd16Ux8( expr_vA, expr_vB ) \
binop(Iop_MullEven16Ux8, \
binop(Iop_ShrV128, expr_vA, mkU8(16)), \
binop(Iop_ShrV128, expr_vB, mkU8(16)))
#define MK_Iop_MullOdd16Sx8( expr_vA, expr_vB ) \
binop(Iop_MullEven16Sx8, \
binop(Iop_ShrV128, expr_vA, mkU8(16)), \
binop(Iop_ShrV128, expr_vB, mkU8(16)))
static IRExpr* /* :: Ity_I64 */ mk64lo32Sto64 ( IRExpr* src )
{
vassert(typeOfIRExpr(irsb->tyenv, src) == Ity_I64);
return unop(Iop_32Sto64, unop(Iop_64to32, src));
}
static IRExpr* /* :: Ity_I64 */ mk64lo32Uto64 ( IRExpr* src )
{
vassert(typeOfIRExpr(irsb->tyenv, src) == Ity_I64);
return unop(Iop_32Uto64, unop(Iop_64to32, src));
}
static IROp mkSzOp ( IRType ty, IROp op8 )
{
Int adj;
vassert(ty == Ity_I8 || ty == Ity_I16 ||
ty == Ity_I32 || ty == Ity_I64);
vassert(op8 == Iop_Add8 || op8 == Iop_Sub8 || op8 == Iop_Mul8 ||
op8 == Iop_Or8 || op8 == Iop_And8 || op8 == Iop_Xor8 ||
op8 == Iop_Shl8 || op8 == Iop_Shr8 || op8 == Iop_Sar8 ||
op8 == Iop_CmpEQ8 || op8 == Iop_CmpNE8 ||
op8 == Iop_Not8 );
adj = ty==Ity_I8 ? 0 : (ty==Ity_I16 ? 1 : (ty==Ity_I32 ? 2 : 3));
return adj + op8;
}
/* Make sure we get valid 32 and 64bit addresses */
static Addr64 mkSzAddr ( IRType ty, Addr64 addr )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ( ty == Ity_I64 ?
(Addr64)addr :
(Addr64)extend_s_32to64( toUInt(addr) ) );
}
/* sz, ULong -> IRExpr */
static IRExpr* mkSzImm ( IRType ty, ULong imm64 )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ty == Ity_I64 ? mkU64(imm64) : mkU32((UInt)imm64);
}
/* sz, ULong -> IRConst */
static IRConst* mkSzConst ( IRType ty, ULong imm64 )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ( ty == Ity_I64 ?
IRConst_U64(imm64) :
IRConst_U32((UInt)imm64) );
}
/* Sign extend imm16 -> IRExpr* */
static IRExpr* mkSzExtendS16 ( IRType ty, UInt imm16 )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ( ty == Ity_I64 ?
mkU64(extend_s_16to64(imm16)) :
mkU32(extend_s_16to32(imm16)) );
}
/* Sign extend imm32 -> IRExpr* */
static IRExpr* mkSzExtendS32 ( IRType ty, UInt imm32 )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ( ty == Ity_I64 ?
mkU64(extend_s_32to64(imm32)) :
mkU32(imm32) );
}
/* IR narrows I32/I64 -> I8/I16/I32 */
static IRExpr* mkNarrowTo8 ( IRType ty, IRExpr* src )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ty == Ity_I64 ? unop(Iop_64to8, src) : unop(Iop_32to8, src);
}
static IRExpr* mkNarrowTo16 ( IRType ty, IRExpr* src )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ty == Ity_I64 ? unop(Iop_64to16, src) : unop(Iop_32to16, src);
}
static IRExpr* mkNarrowTo32 ( IRType ty, IRExpr* src )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
return ty == Ity_I64 ? unop(Iop_64to32, src) : src;
}
/* Signed/Unsigned IR widens I8/I16/I32 -> I32/I64 */
static IRExpr* mkWidenFrom8 ( IRType ty, IRExpr* src, Bool sined )
{
IROp op;
vassert(ty == Ity_I32 || ty == Ity_I64);
if (sined) op = (ty==Ity_I32) ? Iop_8Sto32 : Iop_8Sto64;
else op = (ty==Ity_I32) ? Iop_8Uto32 : Iop_8Uto64;
return unop(op, src);
}
static IRExpr* mkWidenFrom16 ( IRType ty, IRExpr* src, Bool sined )
{
IROp op;
vassert(ty == Ity_I32 || ty == Ity_I64);
if (sined) op = (ty==Ity_I32) ? Iop_16Sto32 : Iop_16Sto64;
else op = (ty==Ity_I32) ? Iop_16Uto32 : Iop_16Uto64;
return unop(op, src);
}
static IRExpr* mkWidenFrom32 ( IRType ty, IRExpr* src, Bool sined )
{
vassert(ty == Ity_I32 || ty == Ity_I64);
if (ty == Ity_I32)
return src;
return (sined) ? unop(Iop_32Sto64, src) : unop(Iop_32Uto64, src);
}
static Int integerGuestRegOffset ( UInt archreg )
{
vassert(archreg < 32);
// jrs: probably not necessary; only matters if we reference sub-parts
// of the ppc registers, but that isn't the case
// later: this might affect Altivec though?
vassert(host_is_bigendian);
switch (archreg) {
case 0: return offsetofPPCGuestState(guest_GPR0);
case 1: return offsetofPPCGuestState(guest_GPR1);
case 2: return offsetofPPCGuestState(guest_GPR2);
case 3: return offsetofPPCGuestState(guest_GPR3);
case 4: return offsetofPPCGuestState(guest_GPR4);
case 5: return offsetofPPCGuestState(guest_GPR5);
case 6: return offsetofPPCGuestState(guest_GPR6);
case 7: return offsetofPPCGuestState(guest_GPR7);
case 8: return offsetofPPCGuestState(guest_GPR8);
case 9: return offsetofPPCGuestState(guest_GPR9);
case 10: return offsetofPPCGuestState(guest_GPR10);
case 11: return offsetofPPCGuestState(guest_GPR11);
case 12: return offsetofPPCGuestState(guest_GPR12);
case 13: return offsetofPPCGuestState(guest_GPR13);
case 14: return offsetofPPCGuestState(guest_GPR14);
case 15: return offsetofPPCGuestState(guest_GPR15);
case 16: return offsetofPPCGuestState(guest_GPR16);
case 17: return offsetofPPCGuestState(guest_GPR17);
case 18: return offsetofPPCGuestState(guest_GPR18);
case 19: return offsetofPPCGuestState(guest_GPR19);
case 20: return offsetofPPCGuestState(guest_GPR20);
case 21: return offsetofPPCGuestState(guest_GPR21);
case 22: return offsetofPPCGuestState(guest_GPR22);
case 23: return offsetofPPCGuestState(guest_GPR23);
case 24: return offsetofPPCGuestState(guest_GPR24);
case 25: return offsetofPPCGuestState(guest_GPR25);
case 26: return offsetofPPCGuestState(guest_GPR26);
case 27: return offsetofPPCGuestState(guest_GPR27);
case 28: return offsetofPPCGuestState(guest_GPR28);
case 29: return offsetofPPCGuestState(guest_GPR29);
case 30: return offsetofPPCGuestState(guest_GPR30);
case 31: return offsetofPPCGuestState(guest_GPR31);
default: break;
}
vpanic("integerGuestRegOffset(ppc,be)"); /*notreached*/
}
static IRExpr* getIReg ( UInt archreg )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(archreg < 32);
return IRExpr_Get( integerGuestRegOffset(archreg), ty );
}
/* Ditto, but write to a reg instead. */
static void putIReg ( UInt archreg, IRExpr* e )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(archreg < 32);
vassert(typeOfIRExpr(irsb->tyenv, e) == ty );
stmt( IRStmt_Put(integerGuestRegOffset(archreg), e) );
}
/* Floating point egisters are mapped to VSX registers[0..31]. */
static Int floatGuestRegOffset ( UInt archreg )
{
vassert(archreg < 32);
switch (archreg) {
case 0: return offsetofPPCGuestState(guest_VSR0);
case 1: return offsetofPPCGuestState(guest_VSR1);
case 2: return offsetofPPCGuestState(guest_VSR2);
case 3: return offsetofPPCGuestState(guest_VSR3);
case 4: return offsetofPPCGuestState(guest_VSR4);
case 5: return offsetofPPCGuestState(guest_VSR5);
case 6: return offsetofPPCGuestState(guest_VSR6);
case 7: return offsetofPPCGuestState(guest_VSR7);
case 8: return offsetofPPCGuestState(guest_VSR8);
case 9: return offsetofPPCGuestState(guest_VSR9);
case 10: return offsetofPPCGuestState(guest_VSR10);
case 11: return offsetofPPCGuestState(guest_VSR11);
case 12: return offsetofPPCGuestState(guest_VSR12);
case 13: return offsetofPPCGuestState(guest_VSR13);
case 14: return offsetofPPCGuestState(guest_VSR14);
case 15: return offsetofPPCGuestState(guest_VSR15);
case 16: return offsetofPPCGuestState(guest_VSR16);
case 17: return offsetofPPCGuestState(guest_VSR17);
case 18: return offsetofPPCGuestState(guest_VSR18);
case 19: return offsetofPPCGuestState(guest_VSR19);
case 20: return offsetofPPCGuestState(guest_VSR20);
case 21: return offsetofPPCGuestState(guest_VSR21);
case 22: return offsetofPPCGuestState(guest_VSR22);
case 23: return offsetofPPCGuestState(guest_VSR23);
case 24: return offsetofPPCGuestState(guest_VSR24);
case 25: return offsetofPPCGuestState(guest_VSR25);
case 26: return offsetofPPCGuestState(guest_VSR26);
case 27: return offsetofPPCGuestState(guest_VSR27);
case 28: return offsetofPPCGuestState(guest_VSR28);
case 29: return offsetofPPCGuestState(guest_VSR29);
case 30: return offsetofPPCGuestState(guest_VSR30);
case 31: return offsetofPPCGuestState(guest_VSR31);
default: break;
}
vpanic("floatGuestRegOffset(ppc)"); /*notreached*/
}
static IRExpr* getFReg ( UInt archreg )
{
vassert(archreg < 32);
return IRExpr_Get( floatGuestRegOffset(archreg), Ity_F64 );
}
/* Ditto, but write to a reg instead. */
static void putFReg ( UInt archreg, IRExpr* e )
{
vassert(archreg < 32);
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_F64);
stmt( IRStmt_Put(floatGuestRegOffset(archreg), e) );
}
/* get Decimal float value. Note, they share floating point register file. */
static IRExpr* getDReg(UInt archreg) {
IRExpr *e;
vassert( archreg < 32 );
e = IRExpr_Get( floatGuestRegOffset( archreg ), Ity_D64 );
return e;
}
/* Read a floating point register pair and combine their contents into a
128-bit value */
static IRExpr *getDReg_pair(UInt archreg) {
IRExpr *high = getDReg( archreg );
IRExpr *low = getDReg( archreg + 1 );
return binop( Iop_D64HLtoD128, high, low );
}
/* Ditto, but write to a reg instead. */
static void putDReg(UInt archreg, IRExpr* e) {
vassert( archreg < 32 );
vassert( typeOfIRExpr(irsb->tyenv, e) == Ity_D64 );
stmt( IRStmt_Put( floatGuestRegOffset( archreg ), e ) );
}
/* Write a 128-bit floating point value into a register pair. */
static void putDReg_pair(UInt archreg, IRExpr *e) {
IRTemp low = newTemp( Ity_D64 );
IRTemp high = newTemp( Ity_D64 );
vassert( archreg < 32 );
vassert( typeOfIRExpr(irsb->tyenv, e) == Ity_D128 );
assign( low, unop( Iop_D128LOtoD64, e ) );
assign( high, unop( Iop_D128HItoD64, e ) );
stmt( IRStmt_Put( floatGuestRegOffset( archreg ), mkexpr( high ) ) );
stmt( IRStmt_Put( floatGuestRegOffset( archreg + 1 ), mkexpr( low ) ) );
}
static Int vsxGuestRegOffset ( UInt archreg )
{
vassert(archreg < 64);
switch (archreg) {
case 0: return offsetofPPCGuestState(guest_VSR0);
case 1: return offsetofPPCGuestState(guest_VSR1);
case 2: return offsetofPPCGuestState(guest_VSR2);
case 3: return offsetofPPCGuestState(guest_VSR3);
case 4: return offsetofPPCGuestState(guest_VSR4);
case 5: return offsetofPPCGuestState(guest_VSR5);
case 6: return offsetofPPCGuestState(guest_VSR6);
case 7: return offsetofPPCGuestState(guest_VSR7);
case 8: return offsetofPPCGuestState(guest_VSR8);
case 9: return offsetofPPCGuestState(guest_VSR9);
case 10: return offsetofPPCGuestState(guest_VSR10);
case 11: return offsetofPPCGuestState(guest_VSR11);
case 12: return offsetofPPCGuestState(guest_VSR12);
case 13: return offsetofPPCGuestState(guest_VSR13);
case 14: return offsetofPPCGuestState(guest_VSR14);
case 15: return offsetofPPCGuestState(guest_VSR15);
case 16: return offsetofPPCGuestState(guest_VSR16);
case 17: return offsetofPPCGuestState(guest_VSR17);
case 18: return offsetofPPCGuestState(guest_VSR18);
case 19: return offsetofPPCGuestState(guest_VSR19);
case 20: return offsetofPPCGuestState(guest_VSR20);
case 21: return offsetofPPCGuestState(guest_VSR21);
case 22: return offsetofPPCGuestState(guest_VSR22);
case 23: return offsetofPPCGuestState(guest_VSR23);
case 24: return offsetofPPCGuestState(guest_VSR24);
case 25: return offsetofPPCGuestState(guest_VSR25);
case 26: return offsetofPPCGuestState(guest_VSR26);
case 27: return offsetofPPCGuestState(guest_VSR27);
case 28: return offsetofPPCGuestState(guest_VSR28);
case 29: return offsetofPPCGuestState(guest_VSR29);
case 30: return offsetofPPCGuestState(guest_VSR30);
case 31: return offsetofPPCGuestState(guest_VSR31);
case 32: return offsetofPPCGuestState(guest_VSR32);
case 33: return offsetofPPCGuestState(guest_VSR33);
case 34: return offsetofPPCGuestState(guest_VSR34);
case 35: return offsetofPPCGuestState(guest_VSR35);
case 36: return offsetofPPCGuestState(guest_VSR36);
case 37: return offsetofPPCGuestState(guest_VSR37);
case 38: return offsetofPPCGuestState(guest_VSR38);
case 39: return offsetofPPCGuestState(guest_VSR39);
case 40: return offsetofPPCGuestState(guest_VSR40);
case 41: return offsetofPPCGuestState(guest_VSR41);
case 42: return offsetofPPCGuestState(guest_VSR42);
case 43: return offsetofPPCGuestState(guest_VSR43);
case 44: return offsetofPPCGuestState(guest_VSR44);
case 45: return offsetofPPCGuestState(guest_VSR45);
case 46: return offsetofPPCGuestState(guest_VSR46);
case 47: return offsetofPPCGuestState(guest_VSR47);
case 48: return offsetofPPCGuestState(guest_VSR48);
case 49: return offsetofPPCGuestState(guest_VSR49);
case 50: return offsetofPPCGuestState(guest_VSR50);
case 51: return offsetofPPCGuestState(guest_VSR51);
case 52: return offsetofPPCGuestState(guest_VSR52);
case 53: return offsetofPPCGuestState(guest_VSR53);
case 54: return offsetofPPCGuestState(guest_VSR54);
case 55: return offsetofPPCGuestState(guest_VSR55);
case 56: return offsetofPPCGuestState(guest_VSR56);
case 57: return offsetofPPCGuestState(guest_VSR57);
case 58: return offsetofPPCGuestState(guest_VSR58);
case 59: return offsetofPPCGuestState(guest_VSR59);
case 60: return offsetofPPCGuestState(guest_VSR60);
case 61: return offsetofPPCGuestState(guest_VSR61);
case 62: return offsetofPPCGuestState(guest_VSR62);
case 63: return offsetofPPCGuestState(guest_VSR63);
default: break;
}
vpanic("vsxGuestRegOffset(ppc)"); /*notreached*/
}
/* Vector registers are mapped to VSX registers[32..63]. */
static Int vectorGuestRegOffset ( UInt archreg )
{
vassert(archreg < 32);
switch (archreg) {
case 0: return offsetofPPCGuestState(guest_VSR32);
case 1: return offsetofPPCGuestState(guest_VSR33);
case 2: return offsetofPPCGuestState(guest_VSR34);
case 3: return offsetofPPCGuestState(guest_VSR35);
case 4: return offsetofPPCGuestState(guest_VSR36);
case 5: return offsetofPPCGuestState(guest_VSR37);
case 6: return offsetofPPCGuestState(guest_VSR38);
case 7: return offsetofPPCGuestState(guest_VSR39);
case 8: return offsetofPPCGuestState(guest_VSR40);
case 9: return offsetofPPCGuestState(guest_VSR41);
case 10: return offsetofPPCGuestState(guest_VSR42);
case 11: return offsetofPPCGuestState(guest_VSR43);
case 12: return offsetofPPCGuestState(guest_VSR44);
case 13: return offsetofPPCGuestState(guest_VSR45);
case 14: return offsetofPPCGuestState(guest_VSR46);
case 15: return offsetofPPCGuestState(guest_VSR47);
case 16: return offsetofPPCGuestState(guest_VSR48);
case 17: return offsetofPPCGuestState(guest_VSR49);
case 18: return offsetofPPCGuestState(guest_VSR50);
case 19: return offsetofPPCGuestState(guest_VSR51);
case 20: return offsetofPPCGuestState(guest_VSR52);
case 21: return offsetofPPCGuestState(guest_VSR53);
case 22: return offsetofPPCGuestState(guest_VSR54);
case 23: return offsetofPPCGuestState(guest_VSR55);
case 24: return offsetofPPCGuestState(guest_VSR56);
case 25: return offsetofPPCGuestState(guest_VSR57);
case 26: return offsetofPPCGuestState(guest_VSR58);
case 27: return offsetofPPCGuestState(guest_VSR59);
case 28: return offsetofPPCGuestState(guest_VSR60);
case 29: return offsetofPPCGuestState(guest_VSR61);
case 30: return offsetofPPCGuestState(guest_VSR62);
case 31: return offsetofPPCGuestState(guest_VSR63);
default: break;
}
vpanic("vextorGuestRegOffset(ppc)"); /*notreached*/
}
static IRExpr* getVReg ( UInt archreg )
{
vassert(archreg < 32);
return IRExpr_Get( vectorGuestRegOffset(archreg), Ity_V128 );
}
/* Ditto, but write to a reg instead. */
static void putVReg ( UInt archreg, IRExpr* e )
{
vassert(archreg < 32);
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_V128);
stmt( IRStmt_Put(vectorGuestRegOffset(archreg), e) );
}
/* Get contents of VSX guest register */
static IRExpr* getVSReg ( UInt archreg )
{
vassert(archreg < 64);
return IRExpr_Get( vsxGuestRegOffset(archreg), Ity_V128 );
}
/* Ditto, but write to a VSX reg instead. */
static void putVSReg ( UInt archreg, IRExpr* e )
{
vassert(archreg < 64);
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_V128);
stmt( IRStmt_Put(vsxGuestRegOffset(archreg), e) );
}
static Int guestCR321offset ( UInt cr )
{
switch (cr) {
case 0: return offsetofPPCGuestState(guest_CR0_321 );
case 1: return offsetofPPCGuestState(guest_CR1_321 );
case 2: return offsetofPPCGuestState(guest_CR2_321 );
case 3: return offsetofPPCGuestState(guest_CR3_321 );
case 4: return offsetofPPCGuestState(guest_CR4_321 );
case 5: return offsetofPPCGuestState(guest_CR5_321 );
case 6: return offsetofPPCGuestState(guest_CR6_321 );
case 7: return offsetofPPCGuestState(guest_CR7_321 );
default: vpanic("guestCR321offset(ppc)");
}
}
static Int guestCR0offset ( UInt cr )
{
switch (cr) {
case 0: return offsetofPPCGuestState(guest_CR0_0 );
case 1: return offsetofPPCGuestState(guest_CR1_0 );
case 2: return offsetofPPCGuestState(guest_CR2_0 );
case 3: return offsetofPPCGuestState(guest_CR3_0 );
case 4: return offsetofPPCGuestState(guest_CR4_0 );
case 5: return offsetofPPCGuestState(guest_CR5_0 );
case 6: return offsetofPPCGuestState(guest_CR6_0 );
case 7: return offsetofPPCGuestState(guest_CR7_0 );
default: vpanic("guestCR3offset(ppc)");
}
}
/* Generate an IR sequence to do a popcount operation on the supplied
IRTemp, and return a new IRTemp holding the result. 'ty' may be
Ity_I32 or Ity_I64 only. */
static IRTemp gen_POPCOUNT ( IRType ty, IRTemp src )
{
Int i, shift[6];
IRTemp mask[6];
IRTemp old = IRTemp_INVALID;
IRTemp nyu = IRTemp_INVALID;
vassert(ty == Ity_I64 || ty == Ity_I32);
if (ty == Ity_I32) {
for (i = 0; i < 5; i++) {
mask[i] = newTemp(ty);
shift[i] = 1 << i;
}
assign(mask[0], mkU32(0x55555555));
assign(mask[1], mkU32(0x33333333));
assign(mask[2], mkU32(0x0F0F0F0F));
assign(mask[3], mkU32(0x00FF00FF));
assign(mask[4], mkU32(0x0000FFFF));
old = src;
for (i = 0; i < 5; i++) {
nyu = newTemp(ty);
assign(nyu,
binop(Iop_Add32,
binop(Iop_And32,
mkexpr(old),
mkexpr(mask[i])),
binop(Iop_And32,
binop(Iop_Shr32, mkexpr(old), mkU8(shift[i])),
mkexpr(mask[i]))));
old = nyu;
}
return nyu;
}
// else, ty == Ity_I64
for (i = 0; i < 6; i++) {
mask[i] = newTemp( Ity_I64 );
shift[i] = 1 << i;
}
assign( mask[0], mkU64( 0x5555555555555555ULL ) );
assign( mask[1], mkU64( 0x3333333333333333ULL ) );
assign( mask[2], mkU64( 0x0F0F0F0F0F0F0F0FULL ) );
assign( mask[3], mkU64( 0x00FF00FF00FF00FFULL ) );
assign( mask[4], mkU64( 0x0000FFFF0000FFFFULL ) );
assign( mask[5], mkU64( 0x00000000FFFFFFFFULL ) );
old = src;
for (i = 0; i < 6; i++) {
nyu = newTemp( Ity_I64 );
assign( nyu,
binop( Iop_Add64,
binop( Iop_And64, mkexpr( old ), mkexpr( mask[i] ) ),
binop( Iop_And64,
binop( Iop_Shr64, mkexpr( old ), mkU8( shift[i] ) ),
mkexpr( mask[i] ) ) ) );
old = nyu;
}
return nyu;
}
// ROTL(src32/64, rot_amt5/6)
static IRExpr* /* :: Ity_I32/64 */ ROTL ( IRExpr* src,
IRExpr* rot_amt )
{
IRExpr *mask, *rot;
vassert(typeOfIRExpr(irsb->tyenv,rot_amt) == Ity_I8);
if (typeOfIRExpr(irsb->tyenv,src) == Ity_I64) {
// rot = (src << rot_amt) | (src >> (64-rot_amt))
mask = binop(Iop_And8, rot_amt, mkU8(63));
rot = binop(Iop_Or64,
binop(Iop_Shl64, src, mask),
binop(Iop_Shr64, src, binop(Iop_Sub8, mkU8(64), mask)));
} else {
// rot = (src << rot_amt) | (src >> (32-rot_amt))
mask = binop(Iop_And8, rot_amt, mkU8(31));
rot = binop(Iop_Or32,
binop(Iop_Shl32, src, mask),
binop(Iop_Shr32, src, binop(Iop_Sub8, mkU8(32), mask)));
}
/* Note: the MuxOX is not merely an optimisation; it's needed
because otherwise the Shr is a shift by the word size when
mask denotes zero. For rotates by immediates, a lot of
this junk gets folded out. */
return IRExpr_Mux0X( mask, /* zero rotate */ src,
/* non-zero rotate */ rot );
}
/* Standard effective address calc: (rA + rB) */
static IRExpr* ea_rA_idxd ( UInt rA, UInt rB )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rA < 32);
vassert(rB < 32);
return binop(mkSzOp(ty, Iop_Add8), getIReg(rA), getIReg(rB));
}
/* Standard effective address calc: (rA + simm) */
static IRExpr* ea_rA_simm ( UInt rA, UInt simm16 )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rA < 32);
return binop(mkSzOp(ty, Iop_Add8), getIReg(rA),
mkSzExtendS16(ty, simm16));
}
/* Standard effective address calc: (rA|0) */
static IRExpr* ea_rAor0 ( UInt rA )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rA < 32);
if (rA == 0) {
return mkSzImm(ty, 0);
} else {
return getIReg(rA);
}
}
/* Standard effective address calc: (rA|0) + rB */
static IRExpr* ea_rAor0_idxd ( UInt rA, UInt rB )
{
vassert(rA < 32);
vassert(rB < 32);
return (rA == 0) ? getIReg(rB) : ea_rA_idxd( rA, rB );
}
/* Standard effective address calc: (rA|0) + simm16 */
static IRExpr* ea_rAor0_simm ( UInt rA, UInt simm16 )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rA < 32);
if (rA == 0) {
return mkSzExtendS16(ty, simm16);
} else {
return ea_rA_simm( rA, simm16 );
}
}
/* Align effective address */
static IRExpr* addr_align( IRExpr* addr, UChar align )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
Long mask;
switch (align) {
case 1: return addr; // byte aligned
case 2: mask = ((Long)-1) << 1; break; // half-word aligned
case 4: mask = ((Long)-1) << 2; break; // word aligned
case 16: mask = ((Long)-1) << 4; break; // quad-word aligned
default:
vex_printf("addr_align: align = %u\n", align);
vpanic("addr_align(ppc)");
}
vassert(typeOfIRExpr(irsb->tyenv,addr) == ty);
return binop( mkSzOp(ty, Iop_And8), addr, mkSzImm(ty, mask) );
}
/* Exit the trace if ADDR (intended to be a guest memory address) is
not ALIGN-aligned, generating a request for a SIGBUS followed by a
restart of the current insn. */
static void gen_SIGBUS_if_misaligned ( IRTemp addr, UChar align )
{
vassert(align == 4 || align == 8);
if (mode64) {
vassert(typeOfIRTemp(irsb->tyenv, addr) == Ity_I64);
stmt(
IRStmt_Exit(
binop(Iop_CmpNE64,
binop(Iop_And64, mkexpr(addr), mkU64(align-1)),
mkU64(0)),
Ijk_SigBUS,
IRConst_U64( guest_CIA_curr_instr ), OFFB_CIA
)
);
} else {
vassert(typeOfIRTemp(irsb->tyenv, addr) == Ity_I32);
stmt(
IRStmt_Exit(
binop(Iop_CmpNE32,
binop(Iop_And32, mkexpr(addr), mkU32(align-1)),
mkU32(0)),
Ijk_SigBUS,
IRConst_U32( guest_CIA_curr_instr ), OFFB_CIA
)
);
}
}
/* Generate AbiHints which mark points at which the ELF or PowerOpen
ABIs say that the stack red zone (viz, -N(r1) .. -1(r1), for some
N) becomes undefined. That is at function calls and returns. ELF
ppc32 doesn't have this "feature" (how fortunate for it). nia is
the address of the next instruction to be executed.
*/
static void make_redzone_AbiHint ( VexAbiInfo* vbi,
IRTemp nia, HChar* who )
{
Int szB = vbi->guest_stack_redzone_size;
if (0) vex_printf("AbiHint: %s\n", who);
vassert(szB >= 0);
if (szB > 0) {
if (mode64) {
vassert(typeOfIRTemp(irsb->tyenv, nia) == Ity_I64);
stmt( IRStmt_AbiHint(
binop(Iop_Sub64, getIReg(1), mkU64(szB)),
szB,
mkexpr(nia)
));
} else {
vassert(typeOfIRTemp(irsb->tyenv, nia) == Ity_I32);
stmt( IRStmt_AbiHint(
binop(Iop_Sub32, getIReg(1), mkU32(szB)),
szB,
mkexpr(nia)
));
}
}
}
/*------------------------------------------------------------*/
/*--- Helpers for condition codes. ---*/
/*------------------------------------------------------------*/
/* Condition register layout.
In the hardware, CR is laid out like this. The leftmost end is the
most significant bit in the register; however the IBM documentation
numbers the bits backwards for some reason.
CR0 CR1 .......... CR6 CR7
0 .. 3 ....................... 28 .. 31 (IBM bit numbering)
31 28 3 0 (normal bit numbering)
Each CR field is 4 bits: [<,>,==,SO]
Hence in IBM's notation, BI=0 is CR7[SO], BI=1 is CR7[==], etc.
Indexing from BI to guest state:
let n = BI / 4
off = BI % 4
this references CR n:
off==0 -> guest_CRn_321 >> 3
off==1 -> guest_CRn_321 >> 2
off==2 -> guest_CRn_321 >> 1
off==3 -> guest_CRn_SO
Bear in mind the only significant bit in guest_CRn_SO is bit 0
(normal notation) and in guest_CRn_321 the significant bits are
3, 2 and 1 (normal notation).
*/
static void putCR321 ( UInt cr, IRExpr* e )
{
vassert(cr < 8);
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
stmt( IRStmt_Put(guestCR321offset(cr), e) );
}
static void putCR0 ( UInt cr, IRExpr* e )
{
vassert(cr < 8);
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
stmt( IRStmt_Put(guestCR0offset(cr), e) );
}
static IRExpr* /* :: Ity_I8 */ getCR0 ( UInt cr )
{
vassert(cr < 8);
return IRExpr_Get(guestCR0offset(cr), Ity_I8);
}
static IRExpr* /* :: Ity_I8 */ getCR321 ( UInt cr )
{
vassert(cr < 8);
return IRExpr_Get(guestCR321offset(cr), Ity_I8);
}
/* Fetch the specified CR bit (as per IBM/hardware notation) and
return it at the bottom of an I32; the top 31 bits are guaranteed
to be zero. */
static IRExpr* /* :: Ity_I32 */ getCRbit ( UInt bi )
{
UInt n = bi / 4;
UInt off = bi % 4;
vassert(bi < 32);
if (off == 3) {
/* Fetch the SO bit for this CR field */
/* Note: And32 is redundant paranoia iff guest state only has 0
or 1 in that slot. */
return binop(Iop_And32, unop(Iop_8Uto32, getCR0(n)), mkU32(1));
} else {
/* Fetch the <, > or == bit for this CR field */
return binop( Iop_And32,
binop( Iop_Shr32,
unop(Iop_8Uto32, getCR321(n)),
mkU8(toUChar(3-off)) ),
mkU32(1) );
}
}
/* Dually, write the least significant bit of BIT to the specified CR
bit. Indexing as per getCRbit. */
static void putCRbit ( UInt bi, IRExpr* bit )
{
UInt n, off;
IRExpr* safe;
vassert(typeOfIRExpr(irsb->tyenv,bit) == Ity_I32);
safe = binop(Iop_And32, bit, mkU32(1));
n = bi / 4;
off = bi % 4;
vassert(bi < 32);
if (off == 3) {
/* This is the SO bit for this CR field */
putCR0(n, unop(Iop_32to8, safe));
} else {
off = 3 - off;
vassert(off == 1 || off == 2 || off == 3);
putCR321(
n,
unop( Iop_32to8,
binop( Iop_Or32,
/* old value with field masked out */
binop(Iop_And32, unop(Iop_8Uto32, getCR321(n)),
mkU32(~(1 << off))),
/* new value in the right place */
binop(Iop_Shl32, safe, mkU8(toUChar(off)))
)
)
);
}
}
/* Fetch the specified CR bit (as per IBM/hardware notation) and
return it somewhere in an I32; it does not matter where, but
whichever bit it is, all other bits are guaranteed to be zero. In
other words, the I32-typed expression will be zero if the bit is
zero and nonzero if the bit is 1. Write into *where the index
of where the bit will be. */
static
IRExpr* /* :: Ity_I32 */ getCRbit_anywhere ( UInt bi, Int* where )
{
UInt n = bi / 4;
UInt off = bi % 4;
vassert(bi < 32);
if (off == 3) {
/* Fetch the SO bit for this CR field */
/* Note: And32 is redundant paranoia iff guest state only has 0
or 1 in that slot. */
*where = 0;
return binop(Iop_And32, unop(Iop_8Uto32, getCR0(n)), mkU32(1));
} else {
/* Fetch the <, > or == bit for this CR field */
*where = 3-off;
return binop( Iop_And32,
unop(Iop_8Uto32, getCR321(n)),
mkU32(1 << (3-off)) );
}
}
/* Set the CR0 flags following an arithmetic operation.
(Condition Register CR0 Field Definition, PPC32 p60)
*/
static IRExpr* getXER_SO ( void );
static void set_CR0 ( IRExpr* result )
{
vassert(typeOfIRExpr(irsb->tyenv,result) == Ity_I32 ||
typeOfIRExpr(irsb->tyenv,result) == Ity_I64);
if (mode64) {
putCR321( 0, unop(Iop_64to8,
binop(Iop_CmpORD64S, result, mkU64(0))) );
} else {
putCR321( 0, unop(Iop_32to8,
binop(Iop_CmpORD32S, result, mkU32(0))) );
}
putCR0( 0, getXER_SO() );
}
/* Set the CR6 flags following an AltiVec compare operation.
* NOTE: This also works for VSX single-precision compares.
* */
static void set_AV_CR6 ( IRExpr* result, Bool test_all_ones )
{
/* CR6[0:3] = {all_ones, 0, all_zeros, 0}
all_ones = (v[0] && v[1] && v[2] && v[3])
all_zeros = ~(v[0] || v[1] || v[2] || v[3])
*/
IRTemp v0 = newTemp(Ity_V128);
IRTemp v1 = newTemp(Ity_V128);
IRTemp v2 = newTemp(Ity_V128);
IRTemp v3 = newTemp(Ity_V128);
IRTemp rOnes = newTemp(Ity_I8);
IRTemp rZeros = newTemp(Ity_I8);
vassert(typeOfIRExpr(irsb->tyenv,result) == Ity_V128);
assign( v0, result );
assign( v1, binop(Iop_ShrV128, result, mkU8(32)) );
assign( v2, binop(Iop_ShrV128, result, mkU8(64)) );
assign( v3, binop(Iop_ShrV128, result, mkU8(96)) );
assign( rZeros, unop(Iop_1Uto8,
binop(Iop_CmpEQ32, mkU32(0xFFFFFFFF),
unop(Iop_Not32,
unop(Iop_V128to32,
binop(Iop_OrV128,
binop(Iop_OrV128, mkexpr(v0), mkexpr(v1)),
binop(Iop_OrV128, mkexpr(v2), mkexpr(v3))))
))) );
if (test_all_ones) {
assign( rOnes, unop(Iop_1Uto8,
binop(Iop_CmpEQ32, mkU32(0xFFFFFFFF),
unop(Iop_V128to32,
binop(Iop_AndV128,
binop(Iop_AndV128, mkexpr(v0), mkexpr(v1)),
binop(Iop_AndV128, mkexpr(v2), mkexpr(v3)))
))) );
putCR321( 6, binop(Iop_Or8,
binop(Iop_Shl8, mkexpr(rOnes), mkU8(3)),
binop(Iop_Shl8, mkexpr(rZeros), mkU8(1))) );
} else {
putCR321( 6, binop(Iop_Shl8, mkexpr(rZeros), mkU8(1)) );
}
putCR0( 6, mkU8(0) );
}
/*------------------------------------------------------------*/
/*--- Helpers for XER flags. ---*/
/*------------------------------------------------------------*/
static void putXER_SO ( IRExpr* e )
{
IRExpr* so;
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
so = binop(Iop_And8, e, mkU8(1));
stmt( IRStmt_Put( OFFB_XER_SO, so ) );
}
static void putXER_OV ( IRExpr* e )
{
IRExpr* ov;
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
ov = binop(Iop_And8, e, mkU8(1));
stmt( IRStmt_Put( OFFB_XER_OV, ov ) );
}
static void putXER_CA ( IRExpr* e )
{
IRExpr* ca;
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
ca = binop(Iop_And8, e, mkU8(1));
stmt( IRStmt_Put( OFFB_XER_CA, ca ) );
}
static void putXER_BC ( IRExpr* e )
{
IRExpr* bc;
vassert(typeOfIRExpr(irsb->tyenv, e) == Ity_I8);
bc = binop(Iop_And8, e, mkU8(0x7F));
stmt( IRStmt_Put( OFFB_XER_BC, bc ) );
}
static IRExpr* /* :: Ity_I8 */ getXER_SO ( void )
{
return IRExpr_Get( OFFB_XER_SO, Ity_I8 );
}
static IRExpr* /* :: Ity_I32 */ getXER_SO32 ( void )
{
return binop( Iop_And32, unop(Iop_8Uto32, getXER_SO()), mkU32(1) );
}
static IRExpr* /* :: Ity_I8 */ getXER_OV ( void )
{
return IRExpr_Get( OFFB_XER_OV, Ity_I8 );
}
static IRExpr* /* :: Ity_I32 */ getXER_OV32 ( void )
{
return binop( Iop_And32, unop(Iop_8Uto32, getXER_OV()), mkU32(1) );
}
static IRExpr* /* :: Ity_I32 */ getXER_CA32 ( void )
{
IRExpr* ca = IRExpr_Get( OFFB_XER_CA, Ity_I8 );
return binop( Iop_And32, unop(Iop_8Uto32, ca ), mkU32(1) );
}
static IRExpr* /* :: Ity_I8 */ getXER_BC ( void )
{
return IRExpr_Get( OFFB_XER_BC, Ity_I8 );
}
static IRExpr* /* :: Ity_I32 */ getXER_BC32 ( void )
{
IRExpr* bc = IRExpr_Get( OFFB_XER_BC, Ity_I8 );
return binop( Iop_And32, unop(Iop_8Uto32, bc), mkU32(0x7F) );
}
/* RES is the result of doing OP on ARGL and ARGR. Set %XER.OV and
%XER.SO accordingly. */
static void set_XER_OV_32( UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR )
{
IRTemp t64;
IRExpr* xer_ov;
vassert(op < PPCG_FLAG_OP_NUMBER);
vassert(typeOfIRExpr(irsb->tyenv,res) == Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv,argL) == Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv,argR) == Ity_I32);
# define INT32_MIN 0x80000000
# define XOR2(_aa,_bb) \
binop(Iop_Xor32,(_aa),(_bb))
# define XOR3(_cc,_dd,_ee) \
binop(Iop_Xor32,binop(Iop_Xor32,(_cc),(_dd)),(_ee))
# define AND3(_ff,_gg,_hh) \
binop(Iop_And32,binop(Iop_And32,(_ff),(_gg)),(_hh))
#define NOT(_jj) \
unop(Iop_Not32, (_jj))
switch (op) {
case /* 0 */ PPCG_FLAG_OP_ADD:
case /* 1 */ PPCG_FLAG_OP_ADDE:
/* (argL^argR^-1) & (argL^res) & (1<<31) ?1:0 */
// i.e. ((both_same_sign) & (sign_changed) & (sign_mask))
xer_ov
= AND3( XOR3(argL,argR,mkU32(-1)),
XOR2(argL,res),
mkU32(INT32_MIN) );
/* xer_ov can only be 0 or 1<<31 */
xer_ov
= binop(Iop_Shr32, xer_ov, mkU8(31) );
break;
case /* 2 */ PPCG_FLAG_OP_DIVW:
/* (argL == INT32_MIN && argR == -1) || argR == 0 */
xer_ov
= mkOR1(
mkAND1(
binop(Iop_CmpEQ32, argL, mkU32(INT32_MIN)),
binop(Iop_CmpEQ32, argR, mkU32(-1))
),
binop(Iop_CmpEQ32, argR, mkU32(0) )
);
xer_ov
= unop(Iop_1Uto32, xer_ov);
break;
case /* 3 */ PPCG_FLAG_OP_DIVWU:
/* argR == 0 */
xer_ov
= unop(Iop_1Uto32, binop(Iop_CmpEQ32, argR, mkU32(0)));
break;
case /* 4 */ PPCG_FLAG_OP_MULLW:
/* OV true if result can't be represented in 32 bits
i.e sHi != sign extension of sLo */
t64 = newTemp(Ity_I64);
assign( t64, binop(Iop_MullS32, argL, argR) );
xer_ov
= binop( Iop_CmpNE32,
unop(Iop_64HIto32, mkexpr(t64)),
binop( Iop_Sar32,
unop(Iop_64to32, mkexpr(t64)),
mkU8(31))
);
xer_ov
= unop(Iop_1Uto32, xer_ov);
break;
case /* 5 */ PPCG_FLAG_OP_NEG:
/* argL == INT32_MIN */
xer_ov
= unop( Iop_1Uto32,
binop(Iop_CmpEQ32, argL, mkU32(INT32_MIN)) );
break;
case /* 6 */ PPCG_FLAG_OP_SUBF:
case /* 7 */ PPCG_FLAG_OP_SUBFC:
case /* 8 */ PPCG_FLAG_OP_SUBFE:
/* ((~argL)^argR^-1) & ((~argL)^res) & (1<<31) ?1:0; */
xer_ov
= AND3( XOR3(NOT(argL),argR,mkU32(-1)),
XOR2(NOT(argL),res),
mkU32(INT32_MIN) );
/* xer_ov can only be 0 or 1<<31 */
xer_ov
= binop(Iop_Shr32, xer_ov, mkU8(31) );
break;
case PPCG_FLAG_OP_DIVWEU:
xer_ov
= binop( Iop_Or32,
unop( Iop_1Uto32, binop( Iop_CmpEQ32, argR, mkU32( 0 ) ) ),
unop( Iop_1Uto32, binop( Iop_CmpLT32U, argR, argL ) ) );
break;
case PPCG_FLAG_OP_DIVWE:
/* If argR == 0 of if the result cannot fit in the 32-bit destination register,
* then OV <- 1. If dest reg is 0 AND both dividend and divisor are non-zero,
* an overflow is implied.
*/
xer_ov = binop( Iop_Or32,
unop( Iop_1Uto32, binop( Iop_CmpEQ32, argR, mkU32( 0 ) ) ),
unop( Iop_1Uto32, mkAND1( binop( Iop_CmpEQ32, res, mkU32( 0 ) ),
mkAND1( binop( Iop_CmpNE32, argL, mkU32( 0 ) ),
binop( Iop_CmpNE32, argR, mkU32( 0 ) ) ) ) ) );
break;
default:
vex_printf("set_XER_OV: op = %u\n", op);
vpanic("set_XER_OV(ppc)");
}
/* xer_ov MUST denote either 0 or 1, no other value allowed */
putXER_OV( unop(Iop_32to8, xer_ov) );
/* Update the summary overflow */
putXER_SO( binop(Iop_Or8, getXER_SO(), getXER_OV()) );
# undef INT32_MIN
# undef AND3
# undef XOR3
# undef XOR2
# undef NOT
}
static void set_XER_OV_64( UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR )
{
IRExpr* xer_ov;
vassert(op < PPCG_FLAG_OP_NUMBER);
vassert(typeOfIRExpr(irsb->tyenv,res) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv,argL) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv,argR) == Ity_I64);
# define INT64_MIN 0x8000000000000000ULL
# define XOR2(_aa,_bb) \
binop(Iop_Xor64,(_aa),(_bb))
# define XOR3(_cc,_dd,_ee) \
binop(Iop_Xor64,binop(Iop_Xor64,(_cc),(_dd)),(_ee))
# define AND3(_ff,_gg,_hh) \
binop(Iop_And64,binop(Iop_And64,(_ff),(_gg)),(_hh))
#define NOT(_jj) \
unop(Iop_Not64, (_jj))
switch (op) {
case /* 0 */ PPCG_FLAG_OP_ADD:
case /* 1 */ PPCG_FLAG_OP_ADDE:
/* (argL^argR^-1) & (argL^res) & (1<<63) ? 1:0 */
// i.e. ((both_same_sign) & (sign_changed) & (sign_mask))
xer_ov
= AND3( XOR3(argL,argR,mkU64(-1)),
XOR2(argL,res),
mkU64(INT64_MIN) );
/* xer_ov can only be 0 or 1<<63 */
xer_ov
= unop(Iop_64to1, binop(Iop_Shr64, xer_ov, mkU8(63)));
break;
case /* 2 */ PPCG_FLAG_OP_DIVW:
/* (argL == INT64_MIN && argR == -1) || argR == 0 */
xer_ov
= mkOR1(
mkAND1(
binop(Iop_CmpEQ64, argL, mkU64(INT64_MIN)),
binop(Iop_CmpEQ64, argR, mkU64(-1))
),
binop(Iop_CmpEQ64, argR, mkU64(0) )
);
break;
case /* 3 */ PPCG_FLAG_OP_DIVWU:
/* argR == 0 */
xer_ov
= binop(Iop_CmpEQ64, argR, mkU64(0));
break;
case /* 4 */ PPCG_FLAG_OP_MULLW: {
/* OV true if result can't be represented in 64 bits
i.e sHi != sign extension of sLo */
xer_ov
= binop( Iop_CmpNE32,
unop(Iop_64HIto32, res),
binop( Iop_Sar32,
unop(Iop_64to32, res),
mkU8(31))
);
break;
}
case /* 5 */ PPCG_FLAG_OP_NEG:
/* argL == INT64_MIN */
xer_ov
= binop(Iop_CmpEQ64, argL, mkU64(INT64_MIN));
break;
case /* 6 */ PPCG_FLAG_OP_SUBF:
case /* 7 */ PPCG_FLAG_OP_SUBFC:
case /* 8 */ PPCG_FLAG_OP_SUBFE:
/* ((~argL)^argR^-1) & ((~argL)^res) & (1<<63) ?1:0; */
xer_ov
= AND3( XOR3(NOT(argL),argR,mkU64(-1)),
XOR2(NOT(argL),res),
mkU64(INT64_MIN) );
/* xer_ov can only be 0 or 1<<63 */
xer_ov
= unop(Iop_64to1, binop(Iop_Shr64, xer_ov, mkU8(63)));
break;
case PPCG_FLAG_OP_DIVDE:
/* If argR == 0, we must set the OV bit. But there's another condition
* where we can get overflow set for divde . . . when the
* result cannot fit in the 64-bit destination register. If dest reg is 0 AND
* both dividend and divisor are non-zero, it implies an overflow.
*/
xer_ov
= mkOR1( binop( Iop_CmpEQ64, argR, mkU64( 0 ) ),
mkAND1( binop( Iop_CmpEQ64, res, mkU64( 0 ) ),
mkAND1( binop( Iop_CmpNE64, argL, mkU64( 0 ) ),
binop( Iop_CmpNE64, argR, mkU64( 0 ) ) ) ) );
break;
case PPCG_FLAG_OP_DIVDEU:
/* If argR == 0 or if argL >= argR, set OV. */
xer_ov = mkOR1( binop( Iop_CmpEQ64, argR, mkU64( 0 ) ),
binop( Iop_CmpLE64U, argR, argL ) );
break;
default:
vex_printf("set_XER_OV: op = %u\n", op);
vpanic("set_XER_OV(ppc64)");
}
/* xer_ov MUST denote either 0 or 1, no other value allowed */
putXER_OV( unop(Iop_1Uto8, xer_ov) );
/* Update the summary overflow */
putXER_SO( binop(Iop_Or8, getXER_SO(), getXER_OV()) );
# undef INT64_MIN
# undef AND3
# undef XOR3
# undef XOR2
# undef NOT
}
static void set_XER_OV ( IRType ty, UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR )
{
if (ty == Ity_I32)
set_XER_OV_32( op, res, argL, argR );
else
set_XER_OV_64( op, res, argL, argR );
}
/* RES is the result of doing OP on ARGL and ARGR with the old %XER.CA
value being OLDCA. Set %XER.CA accordingly. */
static void set_XER_CA_32 ( UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR, IRExpr* oldca )
{
IRExpr* xer_ca;
vassert(op < PPCG_FLAG_OP_NUMBER);
vassert(typeOfIRExpr(irsb->tyenv,res) == Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv,argL) == Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv,argR) == Ity_I32);
vassert(typeOfIRExpr(irsb->tyenv,oldca) == Ity_I32);
/* Incoming oldca is assumed to hold the values 0 or 1 only. This
seems reasonable given that it's always generated by
getXER_CA32(), which masks it accordingly. In any case it being
0 or 1 is an invariant of the ppc guest state representation;
if it has any other value, that invariant has been violated. */
switch (op) {
case /* 0 */ PPCG_FLAG_OP_ADD:
/* res <u argL */
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpLT32U, res, argL));
break;
case /* 1 */ PPCG_FLAG_OP_ADDE:
/* res <u argL || (old_ca==1 && res==argL) */
xer_ca
= mkOR1(
binop(Iop_CmpLT32U, res, argL),
mkAND1(
binop(Iop_CmpEQ32, oldca, mkU32(1)),
binop(Iop_CmpEQ32, res, argL)
)
);
xer_ca
= unop(Iop_1Uto32, xer_ca);
break;
case /* 8 */ PPCG_FLAG_OP_SUBFE:
/* res <u argR || (old_ca==1 && res==argR) */
xer_ca
= mkOR1(
binop(Iop_CmpLT32U, res, argR),
mkAND1(
binop(Iop_CmpEQ32, oldca, mkU32(1)),
binop(Iop_CmpEQ32, res, argR)
)
);
xer_ca
= unop(Iop_1Uto32, xer_ca);
break;
case /* 7 */ PPCG_FLAG_OP_SUBFC:
case /* 9 */ PPCG_FLAG_OP_SUBFI:
/* res <=u argR */
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpLE32U, res, argR));
break;
case /* 10 */ PPCG_FLAG_OP_SRAW:
/* The shift amount is guaranteed to be in 0 .. 63 inclusive.
If it is <= 31, behave like SRAWI; else XER.CA is the sign
bit of argL. */
/* This term valid for shift amount < 32 only */
xer_ca
= binop(
Iop_And32,
binop(Iop_Sar32, argL, mkU8(31)),
binop( Iop_And32,
argL,
binop( Iop_Sub32,
binop(Iop_Shl32, mkU32(1),
unop(Iop_32to8,argR)),
mkU32(1) )
)
);
xer_ca
= IRExpr_Mux0X(
/* shift amt > 31 ? */
unop(Iop_1Uto8, binop(Iop_CmpLT32U, mkU32(31), argR)),
/* no -- be like srawi */
unop(Iop_1Uto32, binop(Iop_CmpNE32, xer_ca, mkU32(0))),
/* yes -- get sign bit of argL */
binop(Iop_Shr32, argL, mkU8(31))
);
break;
case /* 11 */ PPCG_FLAG_OP_SRAWI:
/* xer_ca is 1 iff src was negative and bits_shifted_out !=
0. Since the shift amount is known to be in the range
0 .. 31 inclusive the following seems viable:
xer.ca == 1 iff the following is nonzero:
(argL >>s 31) -- either all 0s or all 1s
& (argL & (1<<argR)-1) -- the stuff shifted out */
xer_ca
= binop(
Iop_And32,
binop(Iop_Sar32, argL, mkU8(31)),
binop( Iop_And32,
argL,
binop( Iop_Sub32,
binop(Iop_Shl32, mkU32(1),
unop(Iop_32to8,argR)),
mkU32(1) )
)
);
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpNE32, xer_ca, mkU32(0)));
break;
default:
vex_printf("set_XER_CA: op = %u\n", op);
vpanic("set_XER_CA(ppc)");
}
/* xer_ca MUST denote either 0 or 1, no other value allowed */
putXER_CA( unop(Iop_32to8, xer_ca) );
}
static void set_XER_CA_64 ( UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR, IRExpr* oldca )
{
IRExpr* xer_ca;
vassert(op < PPCG_FLAG_OP_NUMBER);
vassert(typeOfIRExpr(irsb->tyenv,res) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv,argL) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv,argR) == Ity_I64);
vassert(typeOfIRExpr(irsb->tyenv,oldca) == Ity_I64);
/* Incoming oldca is assumed to hold the values 0 or 1 only. This
seems reasonable given that it's always generated by
getXER_CA32(), which masks it accordingly. In any case it being
0 or 1 is an invariant of the ppc guest state representation;
if it has any other value, that invariant has been violated. */
switch (op) {
case /* 0 */ PPCG_FLAG_OP_ADD:
/* res <u argL */
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpLT64U, res, argL));
break;
case /* 1 */ PPCG_FLAG_OP_ADDE:
/* res <u argL || (old_ca==1 && res==argL) */
xer_ca
= mkOR1(
binop(Iop_CmpLT64U, res, argL),
mkAND1(
binop(Iop_CmpEQ64, oldca, mkU64(1)),
binop(Iop_CmpEQ64, res, argL)
)
);
xer_ca
= unop(Iop_1Uto32, xer_ca);
break;
case /* 8 */ PPCG_FLAG_OP_SUBFE:
/* res <u argR || (old_ca==1 && res==argR) */
xer_ca
= mkOR1(
binop(Iop_CmpLT64U, res, argR),
mkAND1(
binop(Iop_CmpEQ64, oldca, mkU64(1)),
binop(Iop_CmpEQ64, res, argR)
)
);
xer_ca
= unop(Iop_1Uto32, xer_ca);
break;
case /* 7 */ PPCG_FLAG_OP_SUBFC:
case /* 9 */ PPCG_FLAG_OP_SUBFI:
/* res <=u argR */
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpLE64U, res, argR));
break;
case /* 10 */ PPCG_FLAG_OP_SRAW:
/* The shift amount is guaranteed to be in 0 .. 31 inclusive.
If it is <= 31, behave like SRAWI; else XER.CA is the sign
bit of argL. */
/* This term valid for shift amount < 31 only */
xer_ca
= binop(
Iop_And64,
binop(Iop_Sar64, argL, mkU8(31)),
binop( Iop_And64,
argL,
binop( Iop_Sub64,
binop(Iop_Shl64, mkU64(1),
unop(Iop_64to8,argR)),
mkU64(1) )
)
);
xer_ca
= IRExpr_Mux0X(
/* shift amt > 31 ? */
unop(Iop_1Uto8, binop(Iop_CmpLT64U, mkU64(31), argR)),
/* no -- be like srawi */
unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))),
/* yes -- get sign bit of argL */
unop(Iop_64to32, binop(Iop_Shr64, argL, mkU8(63)))
);
break;
case /* 11 */ PPCG_FLAG_OP_SRAWI:
/* xer_ca is 1 iff src was negative and bits_shifted_out != 0.
Since the shift amount is known to be in the range 0 .. 31
inclusive the following seems viable:
xer.ca == 1 iff the following is nonzero:
(argL >>s 31) -- either all 0s or all 1s
& (argL & (1<<argR)-1) -- the stuff shifted out */
xer_ca
= binop(
Iop_And64,
binop(Iop_Sar64, argL, mkU8(31)),
binop( Iop_And64,
argL,
binop( Iop_Sub64,
binop(Iop_Shl64, mkU64(1),
unop(Iop_64to8,argR)),
mkU64(1) )
)
);
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0)));
break;
case /* 12 */ PPCG_FLAG_OP_SRAD:
/* The shift amount is guaranteed to be in 0 .. 63 inclusive.
If it is <= 63, behave like SRADI; else XER.CA is the sign
bit of argL. */
/* This term valid for shift amount < 63 only */
xer_ca
= binop(
Iop_And64,
binop(Iop_Sar64, argL, mkU8(63)),
binop( Iop_And64,
argL,
binop( Iop_Sub64,
binop(Iop_Shl64, mkU64(1),
unop(Iop_64to8,argR)),
mkU64(1) )
)
);
xer_ca
= IRExpr_Mux0X(
/* shift amt > 63 ? */
unop(Iop_1Uto8, binop(Iop_CmpLT64U, mkU64(63), argR)),
/* no -- be like sradi */
unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))),
/* yes -- get sign bit of argL */
unop(Iop_64to32, binop(Iop_Shr64, argL, mkU8(63)))
);
break;
case /* 13 */ PPCG_FLAG_OP_SRADI:
/* xer_ca is 1 iff src was negative and bits_shifted_out != 0.
Since the shift amount is known to be in the range 0 .. 63
inclusive, the following seems viable:
xer.ca == 1 iff the following is nonzero:
(argL >>s 63) -- either all 0s or all 1s
& (argL & (1<<argR)-1) -- the stuff shifted out */
xer_ca
= binop(
Iop_And64,
binop(Iop_Sar64, argL, mkU8(63)),
binop( Iop_And64,
argL,
binop( Iop_Sub64,
binop(Iop_Shl64, mkU64(1),
unop(Iop_64to8,argR)),
mkU64(1) )
)
);
xer_ca
= unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0)));
break;
default:
vex_printf("set_XER_CA: op = %u\n", op);
vpanic("set_XER_CA(ppc64)");
}
/* xer_ca MUST denote either 0 or 1, no other value allowed */
putXER_CA( unop(Iop_32to8, xer_ca) );
}
static void set_XER_CA ( IRType ty, UInt op, IRExpr* res,
IRExpr* argL, IRExpr* argR, IRExpr* oldca )
{
if (ty == Ity_I32)
set_XER_CA_32( op, res, argL, argR, oldca );
else
set_XER_CA_64( op, res, argL, argR, oldca );
}
/*------------------------------------------------------------*/
/*--- Read/write to guest-state --- */
/*------------------------------------------------------------*/
static IRExpr* /* :: Ity_I32/64 */ getGST ( PPC_GST reg )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
switch (reg) {
case PPC_GST_SPRG3_RO:
return IRExpr_Get( OFFB_SPRG3_RO, ty );
case PPC_GST_CIA:
return IRExpr_Get( OFFB_CIA, ty );
case PPC_GST_LR:
return IRExpr_Get( OFFB_LR, ty );
case PPC_GST_CTR:
return IRExpr_Get( OFFB_CTR, ty );
case PPC_GST_VRSAVE:
return IRExpr_Get( OFFB_VRSAVE, Ity_I32 );
case PPC_GST_VSCR:
return binop(Iop_And32, IRExpr_Get( OFFB_VSCR,Ity_I32 ),
mkU32(MASK_VSCR_VALID));
case PPC_GST_CR: {
/* Synthesise the entire CR into a single word. Expensive. */
# define FIELD(_n) \
binop(Iop_Shl32, \
unop(Iop_8Uto32, \
binop(Iop_Or8, \
binop(Iop_And8, getCR321(_n), mkU8(7<<1)), \
binop(Iop_And8, getCR0(_n), mkU8(1)) \
) \
), \
mkU8(4 * (7-(_n))) \
)
return binop(Iop_Or32,
binop(Iop_Or32,
binop(Iop_Or32, FIELD(0), FIELD(1)),
binop(Iop_Or32, FIELD(2), FIELD(3))
),
binop(Iop_Or32,
binop(Iop_Or32, FIELD(4), FIELD(5)),
binop(Iop_Or32, FIELD(6), FIELD(7))
)
);
# undef FIELD
}
case PPC_GST_XER:
return binop(Iop_Or32,
binop(Iop_Or32,
binop( Iop_Shl32, getXER_SO32(), mkU8(31)),
binop( Iop_Shl32, getXER_OV32(), mkU8(30))),
binop(Iop_Or32,
binop( Iop_Shl32, getXER_CA32(), mkU8(29)),
getXER_BC32()));
default:
vex_printf("getGST(ppc): reg = %u", reg);
vpanic("getGST(ppc)");
}
}
/* Get a masked word from the given reg */
static IRExpr* /* ::Ity_I32 */ getGST_masked ( PPC_GST reg, UInt mask )
{
IRTemp val = newTemp(Ity_I32);
vassert( reg < PPC_GST_MAX );
switch (reg) {
case PPC_GST_FPSCR: {
/* Vex-generated code expects the FPSCR to be set as follows:
all exceptions masked, round-to-nearest.
This corresponds to a FPSCR value of 0x0. */
/* In the lower 32 bits of FPSCR, we're only keeping track of
* the binary floating point rounding mode, so if the mask isn't
* asking for this, just return 0x0.
*/
if (mask & MASK_FPSCR_RN) {
assign( val, unop( Iop_8Uto32, IRExpr_Get( OFFB_FPROUND, Ity_I8 ) ) );
} else {
assign( val, mkU32(0x0) );
}
break;
}
default:
vex_printf("getGST_masked(ppc): reg = %u", reg);
vpanic("getGST_masked(ppc)");
}
if (mask != 0xFFFFFFFF) {
return binop(Iop_And32, mkexpr(val), mkU32(mask));
} else {
return mkexpr(val);
}
}
/* Get a masked word from the given reg */
static IRExpr* /* ::Ity_I32 */getGST_masked_upper(PPC_GST reg, ULong mask) {
IRExpr * val;
vassert( reg < PPC_GST_MAX );
switch (reg) {
case PPC_GST_FPSCR: {
/* In the upper 32 bits of FPSCR, we're only keeping track
* of the decimal floating point rounding mode, so if the mask
* isn't asking for this, just return 0x0.
*/
if (mask & MASK_FPSCR_DRN) {
val = binop( Iop_And32,
unop( Iop_8Uto32, IRExpr_Get( OFFB_DFPROUND, Ity_I8 ) ),
unop( Iop_64HIto32, mkU64( mask ) ) );
} else {
val = mkU32( 0x0ULL );
}
break;
}
default:
vex_printf( "getGST_masked_upper(ppc): reg = %u", reg );
vpanic( "getGST_masked_upper(ppc)" );
}
return val;
}
/* Fetch the specified REG[FLD] nibble (as per IBM/hardware notation)
and return it at the bottom of an I32; the top 27 bits are
guaranteed to be zero. */
static IRExpr* /* ::Ity_I32 */ getGST_field ( PPC_GST reg, UInt fld )
{
UInt shft, mask;
vassert( fld < 8 );
vassert( reg < PPC_GST_MAX );
shft = 4*(7-fld);
mask = 0xF<<shft;
switch (reg) {
case PPC_GST_XER:
vassert(fld ==7);
return binop(Iop_Or32,
binop(Iop_Or32,
binop(Iop_Shl32, getXER_SO32(), mkU8(3)),
binop(Iop_Shl32, getXER_OV32(), mkU8(2))),
binop( Iop_Shl32, getXER_CA32(), mkU8(1)));
break;
default:
if (shft == 0)
return getGST_masked( reg, mask );
else
return binop(Iop_Shr32,
getGST_masked( reg, mask ),
mkU8(toUChar( shft )));
}
}
static void putGST ( PPC_GST reg, IRExpr* src )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRType ty_src = typeOfIRExpr(irsb->tyenv,src );
vassert( reg < PPC_GST_MAX );
switch (reg) {
case PPC_GST_IP_AT_SYSCALL:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_IP_AT_SYSCALL, src ) );
break;
case PPC_GST_CIA:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_CIA, src ) );
break;
case PPC_GST_LR:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_LR, src ) );
break;
case PPC_GST_CTR:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_CTR, src ) );
break;
case PPC_GST_VRSAVE:
vassert( ty_src == Ity_I32 );
stmt( IRStmt_Put( OFFB_VRSAVE,src));
break;
case PPC_GST_VSCR:
vassert( ty_src == Ity_I32 );
stmt( IRStmt_Put( OFFB_VSCR,
binop(Iop_And32, src,
mkU32(MASK_VSCR_VALID)) ) );
break;
case PPC_GST_XER:
vassert( ty_src == Ity_I32 );
putXER_SO( unop(Iop_32to8, binop(Iop_Shr32, src, mkU8(31))) );
putXER_OV( unop(Iop_32to8, binop(Iop_Shr32, src, mkU8(30))) );
putXER_CA( unop(Iop_32to8, binop(Iop_Shr32, src, mkU8(29))) );
putXER_BC( unop(Iop_32to8, src) );
break;
case PPC_GST_EMWARN:
vassert( ty_src == Ity_I32 );
stmt( IRStmt_Put( OFFB_EMWARN,src) );
break;
case PPC_GST_TISTART:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_TISTART, src) );
break;
case PPC_GST_TILEN:
vassert( ty_src == ty );
stmt( IRStmt_Put( OFFB_TILEN, src) );
break;
default:
vex_printf("putGST(ppc): reg = %u", reg);
vpanic("putGST(ppc)");
}
}
/* Write masked src to the given reg */
static void putGST_masked ( PPC_GST reg, IRExpr* src, ULong mask )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert( reg < PPC_GST_MAX );
vassert( typeOfIRExpr( irsb->tyenv,src ) == Ity_I64 );
switch (reg) {
case PPC_GST_FPSCR: {
/* Allow writes to either binary or decimal floating point
* Rounding Mode
*/
if (mask & MASK_FPSCR_RN) {
stmt( IRStmt_Put( OFFB_FPROUND,
unop( Iop_32to8,
binop( Iop_And32,
unop( Iop_64to32, src ),
mkU32( MASK_FPSCR_RN & mask ) ) ) ) );
} else if (mask & MASK_FPSCR_DRN) {
stmt( IRStmt_Put( OFFB_DFPROUND,
unop( Iop_32to8,
binop( Iop_And32,
unop( Iop_64HIto32, src ),
mkU32( ( MASK_FPSCR_DRN & mask )
>> 32 ) ) ) ) );
}
/* Give EmWarn for attempted writes to:
- Exception Controls
- Non-IEEE Mode
*/
if (mask & 0xFC) { // Exception Control, Non-IEE mode
VexEmWarn ew = EmWarn_PPCexns;
/* If any of the src::exception_control bits are actually set,
side-exit to the next insn, reporting the warning,
so that Valgrind's dispatcher sees the warning. */
putGST( PPC_GST_EMWARN, mkU32(ew) );
stmt(
IRStmt_Exit(
binop(Iop_CmpNE32, mkU32(ew), mkU32(EmWarn_NONE)),
Ijk_EmWarn,
mkSzConst( ty, nextInsnAddr()), OFFB_CIA ));
}
/* Ignore all other writes */
break;
}
default:
vex_printf("putGST_masked(ppc): reg = %u", reg);
vpanic("putGST_masked(ppc)");
}
}
/* Write the least significant nibble of src to the specified
REG[FLD] (as per IBM/hardware notation). */
static void putGST_field ( PPC_GST reg, IRExpr* src, UInt fld )
{
UInt shft;
ULong mask;
vassert( typeOfIRExpr(irsb->tyenv,src ) == Ity_I32 );
vassert( fld < 16 );
vassert( reg < PPC_GST_MAX );
if (fld < 8)
shft = 4*(7-fld);
else
shft = 4*(15-fld);
mask = 0xF<<shft;
switch (reg) {
case PPC_GST_CR:
putCR0 (fld, binop(Iop_And8, mkU8(1 ), unop(Iop_32to8, src)));
putCR321(fld, binop(Iop_And8, mkU8(7<<1), unop(Iop_32to8, src)));
break;
default:
{
IRExpr * src64 = unop( Iop_32Uto64, src );
if (shft == 0) {
putGST_masked( reg, src64, mask );
} else {
putGST_masked( reg,
binop( Iop_Shl64, src64, mkU8( toUChar( shft ) ) ),
mask );
}
}
}
}
/*------------------------------------------------------------*/
/* Helpers for VSX instructions that do floating point
* operations and need to determine if a src contains a
* special FP value.
*
*------------------------------------------------------------*/
#define NONZERO_FRAC_MASK 0x000fffffffffffffULL
#define FP_FRAC_PART(x) binop( Iop_And64, \
mkexpr( x ), \
mkU64( NONZERO_FRAC_MASK ) )
// Returns exponent part of a single precision floating point as I32
static IRExpr * fp_exp_part_sp(IRTemp src)
{
return binop( Iop_And32,
binop( Iop_Shr32, mkexpr( src ), mkU8( 23 ) ),
mkU32( 0xff ) );
}
// Returns exponent part of floating point as I32
static IRExpr * fp_exp_part(IRTemp src, Bool sp)
{
IRExpr * exp;
if (sp)
return fp_exp_part_sp(src);
if (!mode64)
exp = binop( Iop_And32, binop( Iop_Shr32, unop( Iop_64HIto32,
mkexpr( src ) ),
mkU8( 20 ) ), mkU32( 0x7ff ) );
else
exp = unop( Iop_64to32,
binop( Iop_And64,
binop( Iop_Shr64, mkexpr( src ), mkU8( 52 ) ),
mkU64( 0x7ff ) ) );
return exp;
}
static IRExpr * is_Inf_sp(IRTemp src)
{
IRTemp frac_part = newTemp(Ity_I32);
IRExpr * Inf_exp;
assign( frac_part, binop( Iop_And32, mkexpr(src), mkU32(0x007fffff)) );
Inf_exp = binop( Iop_CmpEQ32, fp_exp_part( src, True /*single precision*/ ), mkU32( 0xff ) );
return mkAND1( Inf_exp, binop( Iop_CmpEQ32, mkexpr( frac_part ), mkU32( 0 ) ) );
}
// Infinity: exp = 7ff and fraction is zero; s = 0/1
static IRExpr * is_Inf(IRTemp src, Bool sp)
{
IRExpr * Inf_exp, * hi32, * low32;
IRTemp frac_part;
if (sp)
return is_Inf_sp(src);
frac_part = newTemp(Ity_I64);
assign( frac_part, FP_FRAC_PART(src) );
Inf_exp = binop( Iop_CmpEQ32, fp_exp_part( src, False /*not single precision*/ ), mkU32( 0x7ff ) );
hi32 = unop( Iop_64HIto32, mkexpr( frac_part ) );
low32 = unop( Iop_64to32, mkexpr( frac_part ) );
return mkAND1( Inf_exp, binop( Iop_CmpEQ32, binop( Iop_Or32, low32, hi32 ),
mkU32( 0 ) ) );
}
static IRExpr * is_Zero_sp(IRTemp src)
{
IRTemp sign_less_part = newTemp(Ity_I32);
assign( sign_less_part, binop( Iop_And32, mkexpr( src ), mkU32( SIGN_MASK32 ) ) );
return binop( Iop_CmpEQ32, mkexpr( sign_less_part ), mkU32( 0 ) );
}
// Zero: exp is zero and fraction is zero; s = 0/1
static IRExpr * is_Zero(IRTemp src, Bool sp)
{
IRExpr * hi32, * low32;
IRTemp sign_less_part;
if (sp)
return is_Zero_sp(src);
sign_less_part = newTemp(Ity_I64);
assign( sign_less_part, binop( Iop_And64, mkexpr( src ), mkU64( SIGN_MASK ) ) );
hi32 = unop( Iop_64HIto32, mkexpr( sign_less_part ) );
low32 = unop( Iop_64to32, mkexpr( sign_less_part ) );
return binop( Iop_CmpEQ32, binop( Iop_Or32, low32, hi32 ),
mkU32( 0 ) );
}
/* SNAN: s = 1/0; exp = 0x7ff; fraction is nonzero, with highest bit '1'
* QNAN: s = 1/0; exp = 0x7ff; fraction is nonzero, with highest bit '0'
* This function returns an IRExpr value of '1' for any type of NaN.
*/
static IRExpr * is_NaN(IRTemp src)
{
IRExpr * NaN_exp, * hi32, * low32;
IRTemp frac_part = newTemp(Ity_I64);
assign( frac_part, FP_FRAC_PART(src) );
hi32 = unop( Iop_64HIto32, mkexpr( frac_part ) );
low32 = unop( Iop_64to32, mkexpr( frac_part ) );
NaN_exp = binop( Iop_CmpEQ32, fp_exp_part( src, False /*not single precision*/ ),
mkU32( 0x7ff ) );
return mkAND1( NaN_exp, binop( Iop_CmpNE32, binop( Iop_Or32, low32, hi32 ),
mkU32( 0 ) ) );
}
/* This function returns an IRExpr value of '1' for any type of NaN.
* The passed 'src' argument is assumed to be Ity_I32.
*/
static IRExpr * is_NaN_32(IRTemp src)
{
#define NONZERO_FRAC_MASK32 0x007fffffULL
#define FP_FRAC_PART32(x) binop( Iop_And32, \
mkexpr( x ), \
mkU32( NONZERO_FRAC_MASK32 ) )
IRExpr * frac_part = FP_FRAC_PART32(src);
IRExpr * exp_part = binop( Iop_And32,
binop( Iop_Shr32, mkexpr( src ), mkU8( 23 ) ),
mkU32( 0x0ff ) );
IRExpr * NaN_exp = binop( Iop_CmpEQ32, exp_part, mkU32( 0xff ) );
return mkAND1( NaN_exp, binop( Iop_CmpNE32, frac_part, mkU32( 0 ) ) );
}
/* This helper function performs the negation part of operations of the form:
* "Negate Multiply-<op>"
* where "<op>" is either "Add" or "Sub".
*
* This function takes one argument -- the floating point intermediate result (converted to
* Ity_I64 via Iop_ReinterpF64asI64) that was obtained from the "Multip-<op>" part of
* the operation described above.
*/
static IRTemp getNegatedResult(IRTemp intermediateResult)
{
ULong signbit_mask = 0x8000000000000000ULL;
IRTemp signbit_32 = newTemp(Ity_I32);
IRTemp resultantSignbit = newTemp(Ity_I1);
IRTemp negatedResult = newTemp(Ity_I64);
assign( signbit_32, binop( Iop_Shr32,
unop( Iop_64HIto32,
binop( Iop_And64, mkexpr( intermediateResult ),
mkU64( signbit_mask ) ) ),
mkU8( 31 ) ) );
/* We negate the signbit if and only if the intermediate result from the
* multiply-<op> was NOT a NaN. This is an XNOR predicate.
*/
assign( resultantSignbit,
unop( Iop_Not1,
binop( Iop_CmpEQ32,
binop( Iop_Xor32,
mkexpr( signbit_32 ),
unop( Iop_1Uto32, is_NaN( intermediateResult ) ) ),
mkU32( 1 ) ) ) );
assign( negatedResult,
binop( Iop_Or64,
binop( Iop_And64,
mkexpr( intermediateResult ),
mkU64( ~signbit_mask ) ),
binop( Iop_32HLto64,
binop( Iop_Shl32,
unop( Iop_1Uto32, mkexpr( resultantSignbit ) ),
mkU8( 31 ) ),
mkU32( 0 ) ) ) );
return negatedResult;
}
/* This helper function performs the negation part of operations of the form:
* "Negate Multiply-<op>"
* where "<op>" is either "Add" or "Sub".
*
* This function takes one argument -- the floating point intermediate result (converted to
* Ity_I32 via Iop_ReinterpF32asI32) that was obtained from the "Multip-<op>" part of
* the operation described above.
*/
static IRTemp getNegatedResult_32(IRTemp intermediateResult)
{
UInt signbit_mask = 0x80000000;
IRTemp signbit_32 = newTemp(Ity_I32);
IRTemp resultantSignbit = newTemp(Ity_I1);
IRTemp negatedResult = newTemp(Ity_I32);
assign( signbit_32, binop( Iop_Shr32,
binop( Iop_And32, mkexpr( intermediateResult ),
mkU32( signbit_mask ) ),
mkU8( 31 ) ) );
/* We negate the signbit if and only if the intermediate result from the
* multiply-<op> was NOT a NaN. This is an XNOR predicate.
*/
assign( resultantSignbit,
unop( Iop_Not1,
binop( Iop_CmpEQ32,
binop( Iop_Xor32,
mkexpr( signbit_32 ),
unop( Iop_1Uto32, is_NaN_32( intermediateResult ) ) ),
mkU32( 1 ) ) ) );
assign( negatedResult,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( intermediateResult ),
mkU32( ~signbit_mask ) ),
binop( Iop_Shl32,
unop( Iop_1Uto32, mkexpr( resultantSignbit ) ),
mkU8( 31 ) ) ) );
return negatedResult;
}
/*------------------------------------------------------------*/
/*--- Integer Instruction Translation --- */
/*------------------------------------------------------------*/
/*
Integer Arithmetic Instructions
*/
static Bool dis_int_arith ( UInt theInstr )
{
/* D-Form, XO-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UChar flag_OE = ifieldBIT10(theInstr);
UInt opc2 = ifieldOPClo9(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
Long simm16 = extend_s_16to64(uimm16);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rA = newTemp(ty);
IRTemp rB = newTemp(ty);
IRTemp rD = newTemp(ty);
Bool do_rc = False;
assign( rA, getIReg(rA_addr) );
assign( rB, getIReg(rB_addr) ); // XO-Form: rD, rA, rB
switch (opc1) {
/* D-Form */
case 0x0C: // addic (Add Immediate Carrying, PPC32 p351
DIP("addic r%u,r%u,%d\n", rD_addr, rA_addr, (Int)simm16);
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
mkSzExtendS16(ty, uimm16) ) );
set_XER_CA( ty, PPCG_FLAG_OP_ADD,
mkexpr(rD), mkexpr(rA), mkSzExtendS16(ty, uimm16),
mkSzImm(ty, 0)/*old xer.ca, which is ignored*/ );
break;
case 0x0D: // addic. (Add Immediate Carrying and Record, PPC32 p352)
DIP("addic. r%u,r%u,%d\n", rD_addr, rA_addr, (Int)simm16);
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
mkSzExtendS16(ty, uimm16) ) );
set_XER_CA( ty, PPCG_FLAG_OP_ADD,
mkexpr(rD), mkexpr(rA), mkSzExtendS16(ty, uimm16),
mkSzImm(ty, 0)/*old xer.ca, which is ignored*/ );
do_rc = True; // Always record to CR
flag_rC = 1;
break;
case 0x0E: // addi (Add Immediate, PPC32 p350)
// li rD,val == addi rD,0,val
// la disp(rA) == addi rD,rA,disp
if ( rA_addr == 0 ) {
DIP("li r%u,%d\n", rD_addr, (Int)simm16);
assign( rD, mkSzExtendS16(ty, uimm16) );
} else {
DIP("addi r%u,r%u,%d\n", rD_addr, rA_addr, (Int)simm16);
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
mkSzExtendS16(ty, uimm16) ) );
}
break;
case 0x0F: // addis (Add Immediate Shifted, PPC32 p353)
// lis rD,val == addis rD,0,val
if ( rA_addr == 0 ) {
DIP("lis r%u,%d\n", rD_addr, (Int)simm16);
assign( rD, mkSzExtendS32(ty, uimm16 << 16) );
} else {
DIP("addis r%u,r%u,0x%x\n", rD_addr, rA_addr, (Int)simm16);
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
mkSzExtendS32(ty, uimm16 << 16) ) );
}
break;
case 0x07: // mulli (Multiply Low Immediate, PPC32 p490)
DIP("mulli r%u,r%u,%d\n", rD_addr, rA_addr, (Int)simm16);
if (mode64)
assign( rD, unop(Iop_128to64,
binop(Iop_MullS64, mkexpr(rA),
mkSzExtendS16(ty, uimm16))) );
else
assign( rD, unop(Iop_64to32,
binop(Iop_MullS32, mkexpr(rA),
mkSzExtendS16(ty, uimm16))) );
break;
case 0x08: // subfic (Subtract from Immediate Carrying, PPC32 p540)
DIP("subfic r%u,r%u,%d\n", rD_addr, rA_addr, (Int)simm16);
// rD = simm16 - rA
assign( rD, binop( mkSzOp(ty, Iop_Sub8),
mkSzExtendS16(ty, uimm16),
mkexpr(rA)) );
set_XER_CA( ty, PPCG_FLAG_OP_SUBFI,
mkexpr(rD), mkexpr(rA), mkSzExtendS16(ty, uimm16),
mkSzImm(ty, 0)/*old xer.ca, which is ignored*/ );
break;
/* XO-Form */
case 0x1F:
do_rc = True; // All below record to CR
switch (opc2) {
case 0x10A: // add (Add, PPC32 p347)
DIP("add%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop( mkSzOp(ty, Iop_Add8),
mkexpr(rA), mkexpr(rB) ) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_ADD,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x00A: // addc (Add Carrying, PPC32 p348)
DIP("addc%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop( mkSzOp(ty, Iop_Add8),
mkexpr(rA), mkexpr(rB)) );
set_XER_CA( ty, PPCG_FLAG_OP_ADD,
mkexpr(rD), mkexpr(rA), mkexpr(rB),
mkSzImm(ty, 0)/*old xer.ca, which is ignored*/ );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_ADD,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x08A: { // adde (Add Extended, PPC32 p349)
IRTemp old_xer_ca = newTemp(ty);
DIP("adde%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = rA + rB + XER[CA]
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
binop( mkSzOp(ty, Iop_Add8),
mkexpr(rB), mkexpr(old_xer_ca))) );
set_XER_CA( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), mkexpr(rB),
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
}
case 0x0EA: { // addme (Add to Minus One Extended, PPC32 p354)
IRTemp old_xer_ca = newTemp(ty);
IRExpr *min_one;
if (rB_addr != 0) {
vex_printf("dis_int_arith(ppc)(addme,rB_addr)\n");
return False;
}
DIP("addme%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = rA + (-1) + XER[CA]
// => Just another form of adde
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
min_one = mkSzImm(ty, (Long)-1);
assign( rD, binop( mkSzOp(ty, Iop_Add8), mkexpr(rA),
binop( mkSzOp(ty, Iop_Add8),
min_one, mkexpr(old_xer_ca)) ));
set_XER_CA( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), min_one,
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), min_one );
}
break;
}
case 0x0CA: { // addze (Add to Zero Extended, PPC32 p355)
IRTemp old_xer_ca = newTemp(ty);
if (rB_addr != 0) {
vex_printf("dis_int_arith(ppc)(addze,rB_addr)\n");
return False;
}
DIP("addze%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = rA + (0) + XER[CA]
// => Just another form of adde
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
assign( rD, binop( mkSzOp(ty, Iop_Add8),
mkexpr(rA), mkexpr(old_xer_ca)) );
set_XER_CA( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), mkSzImm(ty, 0),
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_ADDE,
mkexpr(rD), mkexpr(rA), mkSzImm(ty, 0) );
}
break;
}
case 0x1EB: // divw (Divide Word, PPC32 p388)
DIP("divw%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
/* Note:
XER settings are mode independent, and reflect the
overflow of the low-order 32bit result
CR0[LT|GT|EQ] are undefined if flag_rC && mode64
*/
/* rD[hi32] are undefined: setting them to sign of lo32
- makes set_CR0 happy */
IRExpr* dividend = mk64lo32Sto64( mkexpr(rA) );
IRExpr* divisor = mk64lo32Sto64( mkexpr(rB) );
assign( rD, mk64lo32Uto64( binop(Iop_DivS64, dividend,
divisor) ) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVW,
mkexpr(rD), dividend, divisor );
}
} else {
assign( rD, binop(Iop_DivS32, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVW,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
}
/* Note:
if (0x8000_0000 / -1) or (x / 0)
=> rD=undef, if(flag_rC) CR7=undef, if(flag_OE) XER_OV=1
=> But _no_ exception raised. */
break;
case 0x1CB: // divwu (Divide Word Unsigned, PPC32 p389)
DIP("divwu%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
/* Note:
XER settings are mode independent, and reflect the
overflow of the low-order 32bit result
CR0[LT|GT|EQ] are undefined if flag_rC && mode64
*/
IRExpr* dividend = mk64lo32Uto64( mkexpr(rA) );
IRExpr* divisor = mk64lo32Uto64( mkexpr(rB) );
assign( rD, mk64lo32Uto64( binop(Iop_DivU64, dividend,
divisor) ) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVWU,
mkexpr(rD), dividend, divisor );
}
} else {
assign( rD, binop(Iop_DivU32, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVWU,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
}
/* Note: ditto comment divw, for (x / 0) */
break;
case 0x04B: // mulhw (Multiply High Word, PPC32 p488)
if (flag_OE != 0) {
vex_printf("dis_int_arith(ppc)(mulhw,flag_OE)\n");
return False;
}
DIP("mulhw%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
/* rD[hi32] are undefined: setting them to sign of lo32
- makes set_CR0 happy */
assign( rD, binop(Iop_Sar64,
binop(Iop_Mul64,
mk64lo32Sto64( mkexpr(rA) ),
mk64lo32Sto64( mkexpr(rB) )),
mkU8(32)) );
} else {
assign( rD, unop(Iop_64HIto32,
binop(Iop_MullS32,
mkexpr(rA), mkexpr(rB))) );
}
break;
case 0x00B: // mulhwu (Multiply High Word Unsigned, PPC32 p489)
if (flag_OE != 0) {
vex_printf("dis_int_arith(ppc)(mulhwu,flag_OE)\n");
return False;
}
DIP("mulhwu%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
/* rD[hi32] are undefined: setting them to sign of lo32
- makes set_CR0 happy */
assign( rD, binop(Iop_Sar64,
binop(Iop_Mul64,
mk64lo32Uto64( mkexpr(rA) ),
mk64lo32Uto64( mkexpr(rB) ) ),
mkU8(32)) );
} else {
assign( rD, unop(Iop_64HIto32,
binop(Iop_MullU32,
mkexpr(rA), mkexpr(rB))) );
}
break;
case 0x0EB: // mullw (Multiply Low Word, PPC32 p491)
DIP("mullw%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
/* rD[hi32] are undefined: setting them to sign of lo32
- set_XER_OV() and set_CR0() depend on this */
IRExpr *a = unop(Iop_64to32, mkexpr(rA) );
IRExpr *b = unop(Iop_64to32, mkexpr(rB) );
assign( rD, binop(Iop_MullS32, a, b) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_MULLW,
mkexpr(rD),
unop(Iop_32Uto64, a), unop(Iop_32Uto64, b) );
}
} else {
assign( rD, unop(Iop_64to32,
binop(Iop_MullU32,
mkexpr(rA), mkexpr(rB))) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_MULLW,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
}
break;
case 0x068: // neg (Negate, PPC32 p493)
if (rB_addr != 0) {
vex_printf("dis_int_arith(ppc)(neg,rB_addr)\n");
return False;
}
DIP("neg%s%s r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr);
// rD = (~rA) + 1
assign( rD, binop( mkSzOp(ty, Iop_Add8),
unop( mkSzOp(ty, Iop_Not8), mkexpr(rA) ),
mkSzImm(ty, 1)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_NEG,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x028: // subf (Subtract From, PPC32 p537)
DIP("subf%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = rB - rA
assign( rD, binop( mkSzOp(ty, Iop_Sub8),
mkexpr(rB), mkexpr(rA)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_SUBF,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x008: // subfc (Subtract from Carrying, PPC32 p538)
DIP("subfc%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = rB - rA
assign( rD, binop( mkSzOp(ty, Iop_Sub8),
mkexpr(rB), mkexpr(rA)) );
set_XER_CA( ty, PPCG_FLAG_OP_SUBFC,
mkexpr(rD), mkexpr(rA), mkexpr(rB),
mkSzImm(ty, 0)/*old xer.ca, which is ignored*/ );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_SUBFC,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x088: {// subfe (Subtract from Extended, PPC32 p539)
IRTemp old_xer_ca = newTemp(ty);
DIP("subfe%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
// rD = (log not)rA + rB + XER[CA]
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
assign( rD, binop( mkSzOp(ty, Iop_Add8),
unop( mkSzOp(ty, Iop_Not8), mkexpr(rA)),
binop( mkSzOp(ty, Iop_Add8),
mkexpr(rB), mkexpr(old_xer_ca))) );
set_XER_CA( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), mkexpr(rB),
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
}
case 0x0E8: { // subfme (Subtract from -1 Extended, PPC32 p541)
IRTemp old_xer_ca = newTemp(ty);
IRExpr *min_one;
if (rB_addr != 0) {
vex_printf("dis_int_arith(ppc)(subfme,rB_addr)\n");
return False;
}
DIP("subfme%s%s r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr);
// rD = (log not)rA + (-1) + XER[CA]
// => Just another form of subfe
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
min_one = mkSzImm(ty, (Long)-1);
assign( rD, binop( mkSzOp(ty, Iop_Add8),
unop( mkSzOp(ty, Iop_Not8), mkexpr(rA)),
binop( mkSzOp(ty, Iop_Add8),
min_one, mkexpr(old_xer_ca))) );
set_XER_CA( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), min_one,
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), min_one );
}
break;
}
case 0x0C8: { // subfze (Subtract from Zero Extended, PPC32 p542)
IRTemp old_xer_ca = newTemp(ty);
if (rB_addr != 0) {
vex_printf("dis_int_arith(ppc)(subfze,rB_addr)\n");
return False;
}
DIP("subfze%s%s r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr);
// rD = (log not)rA + (0) + XER[CA]
// => Just another form of subfe
assign( old_xer_ca, mkWidenFrom32(ty, getXER_CA32(), False) );
assign( rD, binop( mkSzOp(ty, Iop_Add8),
unop( mkSzOp(ty, Iop_Not8),
mkexpr(rA)), mkexpr(old_xer_ca)) );
set_XER_CA( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), mkSzImm(ty, 0),
mkexpr(old_xer_ca) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_SUBFE,
mkexpr(rD), mkexpr(rA), mkSzImm(ty, 0) );
}
break;
}
/* 64bit Arithmetic */
case 0x49: // mulhd (Multiply High DWord, PPC64 p539)
if (flag_OE != 0) {
vex_printf("dis_int_arith(ppc)(mulhd,flagOE)\n");
return False;
}
DIP("mulhd%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, unop(Iop_128HIto64,
binop(Iop_MullS64,
mkexpr(rA), mkexpr(rB))) );
break;
case 0x9: // mulhdu (Multiply High DWord Unsigned, PPC64 p540)
if (flag_OE != 0) {
vex_printf("dis_int_arith(ppc)(mulhdu,flagOE)\n");
return False;
}
DIP("mulhdu%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, unop(Iop_128HIto64,
binop(Iop_MullU64,
mkexpr(rA), mkexpr(rB))) );
break;
case 0xE9: // mulld (Multiply Low DWord, PPC64 p543)
DIP("mulld%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop(Iop_Mul64, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_MULLW,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
case 0x1E9: // divd (Divide DWord, PPC64 p419)
DIP("divd%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop(Iop_DivS64, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVW,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
/* Note:
if (0x8000_0000_0000_0000 / -1) or (x / 0)
=> rD=undef, if(flag_rC) CR7=undef, if(flag_OE) XER_OV=1
=> But _no_ exception raised. */
case 0x1C9: // divdu (Divide DWord Unsigned, PPC64 p420)
DIP("divdu%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop(Iop_DivU64, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV( ty, PPCG_FLAG_OP_DIVWU,
mkexpr(rD), mkexpr(rA), mkexpr(rB) );
}
break;
/* Note: ditto comment divd, for (x / 0) */
case 0x18B: // divweu (Divide Word Extended Unsigned)
{
/*
* If (RA) >= (RB), or if an attempt is made to perform the division
* <anything> / 0
* then the contents of register RD are undefined as are (if Rc=1) the contents of
* the LT, GT, and EQ bits of CR Field 0. In these cases, if OE=1 then OV is set
* to 1.
*/
IRTemp res = newTemp(Ity_I32);
IRExpr * dividend, * divisor;
DIP("divweu%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
dividend = unop( Iop_64to32, mkexpr( rA ) );
divisor = unop( Iop_64to32, mkexpr( rB ) );
assign( res, binop( Iop_DivU32E, dividend, divisor ) );
assign( rD, binop( Iop_32HLto64, mkU32( 0 ), mkexpr( res ) ) );
} else {
dividend = mkexpr( rA );
divisor = mkexpr( rB );
assign( res, binop( Iop_DivU32E, dividend, divisor ) );
assign( rD, mkexpr( res) );
}
if (flag_OE) {
set_XER_OV_32( PPCG_FLAG_OP_DIVWEU,
mkexpr(res), dividend, divisor );
}
break;
}
case 0x1AB: // divwe (Divide Word Extended)
{
/*
* If the quotient cannot be represented in 32 bits, or if an
* attempt is made to perform the division
* <anything> / 0
* then the contents of register RD are undefined as are (if
* Rc=1) the contents of the LT, GT, and EQ bits of CR
* Field 0. In these cases, if OE=1 then OV is set to 1.
*/
IRTemp res = newTemp(Ity_I32);
IRExpr * dividend, * divisor;
DIP("divwe%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
if (mode64) {
dividend = unop( Iop_64to32, mkexpr( rA ) );
divisor = unop( Iop_64to32, mkexpr( rB ) );
assign( res, binop( Iop_DivS32E, dividend, divisor ) );
assign( rD, binop( Iop_32HLto64, mkU32( 0 ), mkexpr( res ) ) );
} else {
dividend = mkexpr( rA );
divisor = mkexpr( rB );
assign( res, binop( Iop_DivS32E, dividend, divisor ) );
assign( rD, mkexpr( res) );
}
if (flag_OE) {
set_XER_OV_32( PPCG_FLAG_OP_DIVWE,
mkexpr(res), dividend, divisor );
}
break;
}
case 0x1A9: // divde (Divide Doubleword Extended)
/*
* If the quotient cannot be represented in 64 bits, or if an
* attempt is made to perform the division
* <anything> / 0
* then the contents of register RD are undefined as are (if
* Rc=1) the contents of the LT, GT, and EQ bits of CR
* Field 0. In these cases, if OE=1 then OV is set to 1.
*/
DIP("divde%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop(Iop_DivS64E, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV_64( PPCG_FLAG_OP_DIVDE, mkexpr( rD ),
mkexpr( rA ), mkexpr( rB ) );
}
break;
case 0x189: // divdeuo (Divide Doubleword Extended Unsigned)
// Same CR and OV rules as given for divweu above
DIP("divdeu%s%s r%u,r%u,r%u\n",
flag_OE ? "o" : "", flag_rC ? ".":"",
rD_addr, rA_addr, rB_addr);
assign( rD, binop(Iop_DivU64E, mkexpr(rA), mkexpr(rB)) );
if (flag_OE) {
set_XER_OV_64( PPCG_FLAG_OP_DIVDEU, mkexpr( rD ),
mkexpr( rA ), mkexpr( rB ) );
}
break;
default:
vex_printf("dis_int_arith(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_arith(ppc)(opc1)\n");
return False;
}
putIReg( rD_addr, mkexpr(rD) );
if (do_rc && flag_rC) {
set_CR0( mkexpr(rD) );
}
return True;
}
/*
Integer Compare Instructions
*/
static Bool dis_int_cmp ( UInt theInstr )
{
/* D-Form, X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
UChar b22 = toUChar( IFIELD( theInstr, 22, 1 ) );
UChar flag_L = toUChar( IFIELD( theInstr, 21, 1 ) );
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRExpr *a = getIReg(rA_addr);
IRExpr *b;
if (!mode64 && flag_L==1) { // L==1 invalid for 32 bit.
vex_printf("dis_int_cmp(ppc)(flag_L)\n");
return False;
}
if (b22 != 0) {
vex_printf("dis_int_cmp(ppc)(b22)\n");
return False;
}
switch (opc1) {
case 0x0B: // cmpi (Compare Immediate, PPC32 p368)
DIP("cmpi cr%u,%u,r%u,%d\n", crfD, flag_L, rA_addr,
(Int)extend_s_16to32(uimm16));
b = mkSzExtendS16( ty, uimm16 );
if (flag_L == 1) {
putCR321(crfD, unop(Iop_64to8, binop(Iop_CmpORD64S, a, b)));
} else {
a = mkNarrowTo32( ty, a );
b = mkNarrowTo32( ty, b );
putCR321(crfD, unop(Iop_32to8, binop(Iop_CmpORD32S, a, b)));
}
putCR0( crfD, getXER_SO() );
break;
case 0x0A: // cmpli (Compare Logical Immediate, PPC32 p370)
DIP("cmpli cr%u,%u,r%u,0x%x\n", crfD, flag_L, rA_addr, uimm16);
b = mkSzImm( ty, uimm16 );
if (flag_L == 1) {
putCR321(crfD, unop(Iop_64to8, binop(Iop_CmpORD64U, a, b)));
} else {
a = mkNarrowTo32( ty, a );
b = mkNarrowTo32( ty, b );
putCR321(crfD, unop(Iop_32to8, binop(Iop_CmpORD32U, a, b)));
}
putCR0( crfD, getXER_SO() );
break;
/* X Form */
case 0x1F:
if (b0 != 0) {
vex_printf("dis_int_cmp(ppc)(0x1F,b0)\n");
return False;
}
b = getIReg(rB_addr);
switch (opc2) {
case 0x000: // cmp (Compare, PPC32 p367)
DIP("cmp cr%u,%u,r%u,r%u\n", crfD, flag_L, rA_addr, rB_addr);
/* Comparing a reg with itself produces a result which
doesn't depend on the contents of the reg. Therefore
remove the false dependency, which has been known to cause
memcheck to produce false errors. */
if (rA_addr == rB_addr)
a = b = typeOfIRExpr(irsb->tyenv,a) == Ity_I64
? mkU64(0) : mkU32(0);
if (flag_L == 1) {
putCR321(crfD, unop(Iop_64to8, binop(Iop_CmpORD64S, a, b)));
} else {
a = mkNarrowTo32( ty, a );
b = mkNarrowTo32( ty, b );
putCR321(crfD, unop(Iop_32to8,binop(Iop_CmpORD32S, a, b)));
}
putCR0( crfD, getXER_SO() );
break;
case 0x020: // cmpl (Compare Logical, PPC32 p369)
DIP("cmpl cr%u,%u,r%u,r%u\n", crfD, flag_L, rA_addr, rB_addr);
/* Comparing a reg with itself produces a result which
doesn't depend on the contents of the reg. Therefore
remove the false dependency, which has been known to cause
memcheck to produce false errors. */
if (rA_addr == rB_addr)
a = b = typeOfIRExpr(irsb->tyenv,a) == Ity_I64
? mkU64(0) : mkU32(0);
if (flag_L == 1) {
putCR321(crfD, unop(Iop_64to8, binop(Iop_CmpORD64U, a, b)));
} else {
a = mkNarrowTo32( ty, a );
b = mkNarrowTo32( ty, b );
putCR321(crfD, unop(Iop_32to8, binop(Iop_CmpORD32U, a, b)));
}
putCR0( crfD, getXER_SO() );
break;
default:
vex_printf("dis_int_cmp(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_cmp(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Integer Logical Instructions
*/
static Bool dis_int_logic ( UInt theInstr )
{
/* D-Form, X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rS = newTemp(ty);
IRTemp rA = newTemp(ty);
IRTemp rB = newTemp(ty);
IRExpr* irx;
Bool do_rc = False;
assign( rS, getIReg(rS_addr) );
assign( rB, getIReg(rB_addr) );
switch (opc1) {
case 0x1C: // andi. (AND Immediate, PPC32 p358)
DIP("andi. r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_And8), mkexpr(rS),
mkSzImm(ty, uimm16)) );
do_rc = True; // Always record to CR
flag_rC = 1;
break;
case 0x1D: // andis. (AND Immediate Shifted, PPC32 p359)
DIP("andis r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_And8), mkexpr(rS),
mkSzImm(ty, uimm16 << 16)) );
do_rc = True; // Always record to CR
flag_rC = 1;
break;
case 0x18: // ori (OR Immediate, PPC32 p497)
DIP("ori r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_Or8), mkexpr(rS),
mkSzImm(ty, uimm16)) );
break;
case 0x19: // oris (OR Immediate Shifted, PPC32 p498)
DIP("oris r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_Or8), mkexpr(rS),
mkSzImm(ty, uimm16 << 16)) );
break;
case 0x1A: // xori (XOR Immediate, PPC32 p550)
DIP("xori r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_Xor8), mkexpr(rS),
mkSzImm(ty, uimm16)) );
break;
case 0x1B: // xoris (XOR Immediate Shifted, PPC32 p551)
DIP("xoris r%u,r%u,0x%x\n", rA_addr, rS_addr, uimm16);
assign( rA, binop( mkSzOp(ty, Iop_Xor8), mkexpr(rS),
mkSzImm(ty, uimm16 << 16)) );
break;
/* X Form */
case 0x1F:
do_rc = True; // All below record to CR, except for where we return at case end.
switch (opc2) {
case 0x01C: // and (AND, PPC32 p356)
DIP("and%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign(rA, binop( mkSzOp(ty, Iop_And8),
mkexpr(rS), mkexpr(rB)));
break;
case 0x03C: // andc (AND with Complement, PPC32 p357)
DIP("andc%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign(rA, binop( mkSzOp(ty, Iop_And8), mkexpr(rS),
unop( mkSzOp(ty, Iop_Not8),
mkexpr(rB))));
break;
case 0x01A: { // cntlzw (Count Leading Zeros Word, PPC32 p371)
IRExpr* lo32;
if (rB_addr!=0) {
vex_printf("dis_int_logic(ppc)(cntlzw,rB_addr)\n");
return False;
}
DIP("cntlzw%s r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr);
// mode64: count in low word only
lo32 = mode64 ? unop(Iop_64to32, mkexpr(rS)) : mkexpr(rS);
// Iop_Clz32 undefined for arg==0, so deal with that case:
irx = binop(Iop_CmpNE32, lo32, mkU32(0));
assign(rA, mkWidenFrom32(ty,
IRExpr_Mux0X( unop(Iop_1Uto8, irx),
mkU32(32),
unop(Iop_Clz32, lo32)),
False));
// TODO: alternatively: assign(rA, verbose_Clz32(rS));
break;
}
case 0x11C: // eqv (Equivalent, PPC32 p396)
DIP("eqv%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, unop( mkSzOp(ty, Iop_Not8),
binop( mkSzOp(ty, Iop_Xor8),
mkexpr(rS), mkexpr(rB))) );
break;
case 0x3BA: // extsb (Extend Sign Byte, PPC32 p397
if (rB_addr!=0) {
vex_printf("dis_int_logic(ppc)(extsb,rB_addr)\n");
return False;
}
DIP("extsb%s r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr);
if (mode64)
assign( rA, unop(Iop_8Sto64, unop(Iop_64to8, mkexpr(rS))) );
else
assign( rA, unop(Iop_8Sto32, unop(Iop_32to8, mkexpr(rS))) );
break;
case 0x39A: // extsh (Extend Sign Half Word, PPC32 p398)
if (rB_addr!=0) {
vex_printf("dis_int_logic(ppc)(extsh,rB_addr)\n");
return False;
}
DIP("extsh%s r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr);
if (mode64)
assign( rA, unop(Iop_16Sto64,
unop(Iop_64to16, mkexpr(rS))) );
else
assign( rA, unop(Iop_16Sto32,
unop(Iop_32to16, mkexpr(rS))) );
break;
case 0x1DC: // nand (NAND, PPC32 p492)
DIP("nand%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, unop( mkSzOp(ty, Iop_Not8),
binop( mkSzOp(ty, Iop_And8),
mkexpr(rS), mkexpr(rB))) );
break;
case 0x07C: // nor (NOR, PPC32 p494)
DIP("nor%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, unop( mkSzOp(ty, Iop_Not8),
binop( mkSzOp(ty, Iop_Or8),
mkexpr(rS), mkexpr(rB))) );
break;
case 0x1BC: // or (OR, PPC32 p495)
if ((!flag_rC) && rS_addr == rB_addr) {
DIP("mr r%u,r%u\n", rA_addr, rS_addr);
assign( rA, mkexpr(rS) );
} else {
DIP("or%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, binop( mkSzOp(ty, Iop_Or8),
mkexpr(rS), mkexpr(rB)) );
}
break;
case 0x19C: // orc (OR with Complement, PPC32 p496)
DIP("orc%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, binop( mkSzOp(ty, Iop_Or8), mkexpr(rS),
unop(mkSzOp(ty, Iop_Not8), mkexpr(rB))));
break;
case 0x13C: // xor (XOR, PPC32 p549)
DIP("xor%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
assign( rA, binop( mkSzOp(ty, Iop_Xor8),
mkexpr(rS), mkexpr(rB)) );
break;
/* 64bit Integer Logical Instructions */
case 0x3DA: // extsw (Extend Sign Word, PPC64 p430)
if (rB_addr!=0) {
vex_printf("dis_int_logic(ppc)(extsw,rB_addr)\n");
return False;
}
DIP("extsw%s r%u,r%u\n", flag_rC ? ".":"", rA_addr, rS_addr);
assign(rA, unop(Iop_32Sto64, unop(Iop_64to32, mkexpr(rS))));
break;
case 0x03A: // cntlzd (Count Leading Zeros DWord, PPC64 p401)
if (rB_addr!=0) {
vex_printf("dis_int_logic(ppc)(cntlzd,rB_addr)\n");
return False;
}
DIP("cntlzd%s r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr);
// Iop_Clz64 undefined for arg==0, so deal with that case:
irx = binop(Iop_CmpNE64, mkexpr(rS), mkU64(0));
assign(rA, IRExpr_Mux0X( unop(Iop_1Uto8, irx),
mkU64(64),
unop(Iop_Clz64, mkexpr(rS)) ));
// TODO: alternatively: assign(rA, verbose_Clz64(rS));
break;
case 0x1FC: // cmpb (Power6: compare bytes)
DIP("cmpb r%u,r%u,r%u\n", rA_addr, rS_addr, rB_addr);
if (mode64)
assign( rA, unop( Iop_V128to64,
binop( Iop_CmpEQ8x16,
binop( Iop_64HLtoV128, mkU64(0), mkexpr(rS) ),
binop( Iop_64HLtoV128, mkU64(0), mkexpr(rB) )
)) );
else
assign( rA, unop( Iop_V128to32,
binop( Iop_CmpEQ8x16,
unop( Iop_32UtoV128, mkexpr(rS) ),
unop( Iop_32UtoV128, mkexpr(rB) )
)) );
break;
case 0x2DF: { // mftgpr (move floating-point to general purpose register)
IRTemp frB = newTemp(Ity_F64);
DIP("mftgpr r%u,fr%u\n", rS_addr, rB_addr);
assign( frB, getFReg(rB_addr)); // always F64
if (mode64)
assign( rA, unop( Iop_ReinterpF64asI64, mkexpr(frB)) );
else
assign( rA, unop( Iop_64to32, unop( Iop_ReinterpF64asI64, mkexpr(frB))) );
putIReg( rS_addr, mkexpr(rA));
return True;
}
case 0x25F: { // mffgpr (move floating-point from general purpose register)
IRTemp frA = newTemp(Ity_F64);
DIP("mffgpr fr%u,r%u\n", rS_addr, rB_addr);
if (mode64)
assign( frA, unop( Iop_ReinterpI64asF64, mkexpr(rB)) );
else
assign( frA, unop( Iop_ReinterpI64asF64, unop( Iop_32Uto64, mkexpr(rB))) );
putFReg( rS_addr, mkexpr(frA));
return True;
}
case 0x1FA: // popcntd (population count doubleword
{
DIP("popcntd r%u,r%u\n", rA_addr, rS_addr);
IRTemp result = gen_POPCOUNT(ty, rS);
putIReg( rA_addr, mkexpr(result) );
return True;
}
case 0x17A: // popcntw (Population Count Words)
{
DIP("popcntw r%u,r%u\n", rA_addr, rS_addr);
if (mode64) {
IRTemp resultHi, resultLo;
IRTemp argLo = newTemp(Ity_I32);
IRTemp argHi = newTemp(Ity_I32);
assign(argLo, unop(Iop_64to32, mkexpr(rS)));
assign(argHi, unop(Iop_64HIto32, mkexpr(rS)));
resultLo = gen_POPCOUNT(Ity_I32, argLo);
resultHi = gen_POPCOUNT(Ity_I32, argHi);
putIReg( rA_addr, binop(Iop_32HLto64, mkexpr(resultHi), mkexpr(resultLo)));
} else {
IRTemp result = gen_POPCOUNT(ty, rS);
putIReg( rA_addr, mkexpr(result) );
}
return True;
}
case 0x0FC: // bpermd (Bit Permute Doubleword)
{
/* This is a lot of rigmarole to emulate bpermd like this, as it
* could be done much faster by implementing a call to the native
* instruction. However, where possible I want to avoid using new
* native instructions so that we can use valgrind to emulate those
* instructions on older PPC64 hardware.
*/
#define BPERMD_IDX_MASK 0x00000000000000FFULL
#define BPERMD_BIT_MASK 0x8000000000000000ULL
int i;
IRExpr * rS_expr = mkexpr(rS);
IRExpr * res = binop(Iop_And64, mkU64(0), mkU64(0));
DIP("bpermd r%u,r%u,r%u\n", rA_addr, rS_addr, rB_addr);
for (i = 0; i < 8; i++) {
IRTemp idx_tmp = newTemp( Ity_I64 );
IRTemp perm_bit = newTemp( Ity_I64 );
IRTemp idx = newTemp( Ity_I8 );
IRTemp idx_LT64 = newTemp( Ity_I1 );
IRTemp idx_LT64_ity64 = newTemp( Ity_I64 );
assign( idx_tmp,
binop( Iop_And64, mkU64( BPERMD_IDX_MASK ), rS_expr ) );
assign( idx_LT64,
binop( Iop_CmpLT64U, mkexpr( idx_tmp ), mkU64( 64 ) ) );
assign( idx,
binop( Iop_And8,
unop( Iop_1Sto8,
mkexpr(idx_LT64) ),
unop( Iop_64to8, mkexpr( idx_tmp ) ) ) );
/* If idx_LT64 == 0, we must force the perm bit to '0'. Below, we se idx
* to determine which bit of rB to use for the perm bit, and then we shift
* that bit to the MSB position. We AND that with a 64-bit-ized idx_LT64
* to set the final perm bit.
*/
assign( idx_LT64_ity64,
unop( Iop_32Uto64, unop( Iop_1Uto32, mkexpr(idx_LT64 ) ) ) );
assign( perm_bit,
binop( Iop_And64,
mkexpr( idx_LT64_ity64 ),
binop( Iop_Shr64,
binop( Iop_And64,
mkU64( BPERMD_BIT_MASK ),
binop( Iop_Shl64,
mkexpr( rB ),
mkexpr( idx ) ) ),
mkU8( 63 ) ) ) );
res = binop( Iop_Or64,
res,
binop( Iop_Shl64,
mkexpr( perm_bit ),
mkU8( i ) ) );
rS_expr = binop( Iop_Shr64, rS_expr, mkU8( 8 ) );
}
putIReg(rA_addr, res);
return True;
}
default:
vex_printf("dis_int_logic(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_logic(ppc)(opc1)\n");
return False;
}
putIReg( rA_addr, mkexpr(rA) );
if (do_rc && flag_rC) {
set_CR0( mkexpr(rA) );
}
return True;
}
/*
Integer Parity Instructions
*/
static Bool dis_int_parity ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rS = newTemp(ty);
IRTemp rA = newTemp(ty);
IRTemp iTot1 = newTemp(Ity_I32);
IRTemp iTot2 = newTemp(Ity_I32);
IRTemp iTot3 = newTemp(Ity_I32);
IRTemp iTot4 = newTemp(Ity_I32);
IRTemp iTot5 = newTemp(Ity_I32);
IRTemp iTot6 = newTemp(Ity_I32);
IRTemp iTot7 = newTemp(Ity_I32);
IRTemp iTot8 = newTemp(Ity_I32);
IRTemp rS1 = newTemp(ty);
IRTemp rS2 = newTemp(ty);
IRTemp rS3 = newTemp(ty);
IRTemp rS4 = newTemp(ty);
IRTemp rS5 = newTemp(ty);
IRTemp rS6 = newTemp(ty);
IRTemp rS7 = newTemp(ty);
IRTemp iHi = newTemp(Ity_I32);
IRTemp iLo = newTemp(Ity_I32);
IROp to_bit = (mode64 ? Iop_64to1 : Iop_32to1);
IROp shr_op = (mode64 ? Iop_Shr64 : Iop_Shr32);
if (opc1 != 0x1f || rB_addr || b0) {
vex_printf("dis_int_parity(ppc)(0x1F,opc1:rB|b0)\n");
return False;
}
assign( rS, getIReg(rS_addr) );
switch (opc2) {
case 0xba: // prtyd (Parity Doubleword, ISA 2.05 p320)
DIP("prtyd r%u,r%u\n", rA_addr, rS_addr);
assign( iTot1, unop(Iop_1Uto32, unop(to_bit, mkexpr(rS))) );
assign( rS1, binop(shr_op, mkexpr(rS), mkU8(8)) );
assign( iTot2, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS1))),
mkexpr(iTot1)) );
assign( rS2, binop(shr_op, mkexpr(rS1), mkU8(8)) );
assign( iTot3, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS2))),
mkexpr(iTot2)) );
assign( rS3, binop(shr_op, mkexpr(rS2), mkU8(8)) );
assign( iTot4, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS3))),
mkexpr(iTot3)) );
if (mode64) {
assign( rS4, binop(shr_op, mkexpr(rS3), mkU8(8)) );
assign( iTot5, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS4))),
mkexpr(iTot4)) );
assign( rS5, binop(shr_op, mkexpr(rS4), mkU8(8)) );
assign( iTot6, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS5))),
mkexpr(iTot5)) );
assign( rS6, binop(shr_op, mkexpr(rS5), mkU8(8)) );
assign( iTot7, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS6))),
mkexpr(iTot6)) );
assign( rS7, binop(shr_op, mkexpr(rS6), mkU8(8)) );
assign( iTot8, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS7))),
mkexpr(iTot7)) );
assign( rA, unop(Iop_32Uto64,
binop(Iop_And32, mkexpr(iTot8), mkU32(1))) );
} else
assign( rA, mkexpr(iTot4) );
break;
case 0x9a: // prtyw (Parity Word, ISA 2.05 p320)
assign( iTot1, unop(Iop_1Uto32, unop(to_bit, mkexpr(rS))) );
assign( rS1, binop(shr_op, mkexpr(rS), mkU8(8)) );
assign( iTot2, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS1))),
mkexpr(iTot1)) );
assign( rS2, binop(shr_op, mkexpr(rS1), mkU8(8)) );
assign( iTot3, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS2))),
mkexpr(iTot2)) );
assign( rS3, binop(shr_op, mkexpr(rS2), mkU8(8)) );
assign( iTot4, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS3))),
mkexpr(iTot3)) );
assign( iLo, unop(Iop_1Uto32, unop(Iop_32to1, mkexpr(iTot4) )) );
if (mode64) {
assign( rS4, binop(shr_op, mkexpr(rS3), mkU8(8)) );
assign( iTot5, unop(Iop_1Uto32, unop(to_bit, mkexpr(rS4))) );
assign( rS5, binop(shr_op, mkexpr(rS4), mkU8(8)) );
assign( iTot6, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS5))),
mkexpr(iTot5)) );
assign( rS6, binop(shr_op, mkexpr(rS5), mkU8(8)) );
assign( iTot7, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS6))),
mkexpr(iTot6)) );
assign( rS7, binop(shr_op, mkexpr(rS6), mkU8(8)));
assign( iTot8, binop(Iop_Add32,
unop(Iop_1Uto32, unop(to_bit, mkexpr(rS7))),
mkexpr(iTot7)) );
assign( iHi, binop(Iop_And32, mkU32(1), mkexpr(iTot8)) ),
assign( rA, binop(Iop_32HLto64, mkexpr(iHi), mkexpr(iLo)) );
} else
assign( rA, binop(Iop_Or32, mkU32(0), mkexpr(iLo)) );
break;
default:
vex_printf("dis_int_parity(ppc)(opc2)\n");
return False;
}
putIReg( rA_addr, mkexpr(rA) );
return True;
}
/*
Integer Rotate Instructions
*/
static Bool dis_int_rot ( UInt theInstr )
{
/* M-Form, MDS-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UChar sh_imm = rB_addr;
UChar MaskBeg = toUChar( IFIELD( theInstr, 6, 5 ) );
UChar MaskEnd = toUChar( IFIELD( theInstr, 1, 5 ) );
UChar msk_imm = toUChar( IFIELD( theInstr, 5, 6 ) );
UChar opc2 = toUChar( IFIELD( theInstr, 2, 3 ) );
UChar b1 = ifieldBIT1(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rS = newTemp(ty);
IRTemp rA = newTemp(ty);
IRTemp rB = newTemp(ty);
IRTemp rot = newTemp(ty);
IRExpr *r;
UInt mask32;
ULong mask64;
assign( rS, getIReg(rS_addr) );
assign( rB, getIReg(rB_addr) );
switch (opc1) {
case 0x14: {
// rlwimi (Rotate Left Word Imm then Mask Insert, PPC32 p500)
DIP("rlwimi%s r%u,r%u,%d,%d,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, MaskBeg, MaskEnd);
if (mode64) {
// tmp32 = (ROTL(rS_Lo32, Imm)
// rA = ((tmp32 || tmp32) & mask64) | (rA & ~mask64)
mask64 = MASK64(31-MaskEnd, 31-MaskBeg);
r = ROTL( unop(Iop_64to32, mkexpr(rS) ), mkU8(sh_imm) );
r = unop(Iop_32Uto64, r);
assign( rot, binop(Iop_Or64, r,
binop(Iop_Shl64, r, mkU8(32))) );
assign( rA,
binop(Iop_Or64,
binop(Iop_And64, mkexpr(rot), mkU64(mask64)),
binop(Iop_And64, getIReg(rA_addr), mkU64(~mask64))) );
}
else {
// rA = (ROTL(rS, Imm) & mask) | (rA & ~mask);
mask32 = MASK32(31-MaskEnd, 31-MaskBeg);
r = ROTL(mkexpr(rS), mkU8(sh_imm));
assign( rA,
binop(Iop_Or32,
binop(Iop_And32, mkU32(mask32), r),
binop(Iop_And32, getIReg(rA_addr), mkU32(~mask32))) );
}
break;
}
case 0x15: {
// rlwinm (Rotate Left Word Imm then AND with Mask, PPC32 p501)
vassert(MaskBeg < 32);
vassert(MaskEnd < 32);
vassert(sh_imm < 32);
if (mode64) {
IRTemp rTmp = newTemp(Ity_I64);
mask64 = MASK64(31-MaskEnd, 31-MaskBeg);
DIP("rlwinm%s r%u,r%u,%d,%d,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, MaskBeg, MaskEnd);
// tmp32 = (ROTL(rS_Lo32, Imm)
// rA = ((tmp32 || tmp32) & mask64)
r = ROTL( unop(Iop_64to32, mkexpr(rS) ), mkU8(sh_imm) );
r = unop(Iop_32Uto64, r);
assign( rTmp, r );
r = NULL;
assign( rot, binop(Iop_Or64, mkexpr(rTmp),
binop(Iop_Shl64, mkexpr(rTmp), mkU8(32))) );
assign( rA, binop(Iop_And64, mkexpr(rot), mkU64(mask64)) );
}
else {
if (MaskBeg == 0 && sh_imm+MaskEnd == 31) {
/* Special-case the ,n,0,31-n form as that is just n-bit
shift left, PPC32 p501 */
DIP("slwi%s r%u,r%u,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm);
assign( rA, binop(Iop_Shl32, mkexpr(rS), mkU8(sh_imm)) );
}
else if (MaskEnd == 31 && sh_imm+MaskBeg == 32) {
/* Special-case the ,32-n,n,31 form as that is just n-bit
unsigned shift right, PPC32 p501 */
DIP("srwi%s r%u,r%u,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, MaskBeg);
assign( rA, binop(Iop_Shr32, mkexpr(rS), mkU8(MaskBeg)) );
}
else {
/* General case. */
mask32 = MASK32(31-MaskEnd, 31-MaskBeg);
DIP("rlwinm%s r%u,r%u,%d,%d,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, MaskBeg, MaskEnd);
// rA = ROTL(rS, Imm) & mask
assign( rA, binop(Iop_And32,
ROTL(mkexpr(rS), mkU8(sh_imm)),
mkU32(mask32)) );
}
}
break;
}
case 0x17: {
// rlwnm (Rotate Left Word then AND with Mask, PPC32 p503
DIP("rlwnm%s r%u,r%u,r%u,%d,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr, MaskBeg, MaskEnd);
if (mode64) {
mask64 = MASK64(31-MaskEnd, 31-MaskBeg);
/* weird insn alert!
tmp32 = (ROTL(rS_Lo32, rB[0-4])
rA = ((tmp32 || tmp32) & mask64)
*/
// note, ROTL does the masking, so we don't do it here
r = ROTL( unop(Iop_64to32, mkexpr(rS)),
unop(Iop_64to8, mkexpr(rB)) );
r = unop(Iop_32Uto64, r);
assign(rot, binop(Iop_Or64, r, binop(Iop_Shl64, r, mkU8(32))));
assign( rA, binop(Iop_And64, mkexpr(rot), mkU64(mask64)) );
} else {
mask32 = MASK32(31-MaskEnd, 31-MaskBeg);
// rA = ROTL(rS, rB[0-4]) & mask
// note, ROTL does the masking, so we don't do it here
assign( rA, binop(Iop_And32,
ROTL(mkexpr(rS),
unop(Iop_32to8, mkexpr(rB))),
mkU32(mask32)) );
}
break;
}
/* 64bit Integer Rotates */
case 0x1E: {
msk_imm = ((msk_imm & 1) << 5) | (msk_imm >> 1);
sh_imm |= b1 << 5;
vassert( msk_imm < 64 );
vassert( sh_imm < 64 );
switch (opc2) {
case 0x4: {
/* r = ROTL64( rS, rB_lo6) */
r = ROTL( mkexpr(rS), unop(Iop_64to8, mkexpr(rB)) );
if (b1 == 0) { // rldcl (Rotl DWord, Clear Left, PPC64 p555)
DIP("rldcl%s r%u,r%u,r%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr, msk_imm);
// note, ROTL does the masking, so we don't do it here
mask64 = MASK64(0, 63-msk_imm);
assign( rA, binop(Iop_And64, r, mkU64(mask64)) );
break;
} else { // rldcr (Rotl DWord, Clear Right, PPC64 p556)
DIP("rldcr%s r%u,r%u,r%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr, msk_imm);
mask64 = MASK64(63-msk_imm, 63);
assign( rA, binop(Iop_And64, r, mkU64(mask64)) );
break;
}
break;
}
case 0x2: // rldic (Rotl DWord Imm, Clear, PPC64 p557)
DIP("rldic%s r%u,r%u,%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, msk_imm);
r = ROTL(mkexpr(rS), mkU8(sh_imm));
mask64 = MASK64(sh_imm, 63-msk_imm);
assign( rA, binop(Iop_And64, r, mkU64(mask64)) );
break;
// later: deal with special case: (msk_imm==0) => SHL(sh_imm)
/*
Hmm... looks like this'll do the job more simply:
r = SHL(rS, sh_imm)
m = ~(1 << (63-msk_imm))
assign(rA, r & m);
*/
case 0x0: // rldicl (Rotl DWord Imm, Clear Left, PPC64 p558)
if (mode64
&& sh_imm + msk_imm == 64 && msk_imm >= 1 && msk_imm <= 63) {
/* special-case the ,64-n,n form as that is just
unsigned shift-right by n */
DIP("srdi%s r%u,r%u,%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, msk_imm);
assign( rA, binop(Iop_Shr64, mkexpr(rS), mkU8(msk_imm)) );
} else {
DIP("rldicl%s r%u,r%u,%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, msk_imm);
r = ROTL(mkexpr(rS), mkU8(sh_imm));
mask64 = MASK64(0, 63-msk_imm);
assign( rA, binop(Iop_And64, r, mkU64(mask64)) );
}
break;
case 0x1: // rldicr (Rotl DWord Imm, Clear Right, PPC64 p559)
if (mode64
&& sh_imm + msk_imm == 63 && sh_imm >= 1 && sh_imm <= 63) {
/* special-case the ,n,63-n form as that is just
shift-left by n */
DIP("sldi%s r%u,r%u,%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, sh_imm);
assign( rA, binop(Iop_Shl64, mkexpr(rS), mkU8(sh_imm)) );
} else {
DIP("rldicr%s r%u,r%u,%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, msk_imm);
r = ROTL(mkexpr(rS), mkU8(sh_imm));
mask64 = MASK64(63-msk_imm, 63);
assign( rA, binop(Iop_And64, r, mkU64(mask64)) );
}
break;
case 0x3: { // rldimi (Rotl DWord Imm, Mask Insert, PPC64 p560)
IRTemp rA_orig = newTemp(ty);
DIP("rldimi%s r%u,r%u,%u,%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm, msk_imm);
r = ROTL(mkexpr(rS), mkU8(sh_imm));
mask64 = MASK64(sh_imm, 63-msk_imm);
assign( rA_orig, getIReg(rA_addr) );
assign( rA, binop(Iop_Or64,
binop(Iop_And64, mkU64(mask64), r),
binop(Iop_And64, mkU64(~mask64),
mkexpr(rA_orig))) );
break;
}
default:
vex_printf("dis_int_rot(ppc)(opc2)\n");
return False;
}
break;
}
default:
vex_printf("dis_int_rot(ppc)(opc1)\n");
return False;
}
putIReg( rA_addr, mkexpr(rA) );
if (flag_rC) {
set_CR0( mkexpr(rA) );
}
return True;
}
/*
Integer Load Instructions
*/
static Bool dis_int_load ( UInt theInstr )
{
/* D-Form, X-Form, DS-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b1 = ifieldBIT1(theInstr);
UChar b0 = ifieldBIT0(theInstr);
Int simm16 = extend_s_16to32(uimm16);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRExpr* val;
switch (opc1) {
case 0x1F: // register offset
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
break;
case 0x3A: // immediate offset: 64bit: ld/ldu/lwa: mask off
// lowest 2 bits of immediate before forming EA
simm16 = simm16 & 0xFFFFFFFC;
default: // immediate offset
assign( EA, ea_rAor0_simm( rA_addr, simm16 ) );
break;
}
switch (opc1) {
case 0x22: // lbz (Load B & Zero, PPC32 p433)
DIP("lbz r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I8, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom8(ty, val, False) );
break;
case 0x23: // lbzu (Load B & Zero, Update, PPC32 p434)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lbzu,rA_addr|rD_addr)\n");
return False;
}
DIP("lbzu r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I8, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom8(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x2A: // lha (Load HW Alg, PPC32 p445)
DIP("lha r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, True) );
break;
case 0x2B: // lhau (Load HW Alg, Update, PPC32 p446)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lhau,rA_addr|rD_addr)\n");
return False;
}
DIP("lhau r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, True) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x28: // lhz (Load HW & Zero, PPC32 p450)
DIP("lhz r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, False) );
break;
case 0x29: // lhzu (Load HW & and Zero, Update, PPC32 p451)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lhzu,rA_addr|rD_addr)\n");
return False;
}
DIP("lhzu r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x20: // lwz (Load W & Zero, PPC32 p460)
DIP("lwz r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I32, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom32(ty, val, False) );
break;
case 0x21: // lwzu (Load W & Zero, Update, PPC32 p461))
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lwzu,rA_addr|rD_addr)\n");
return False;
}
DIP("lwzu r%u,%d(r%u)\n", rD_addr, (Int)simm16, rA_addr);
val = loadBE(Ity_I32, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom32(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
/* X Form */
case 0x1F:
if (b0 != 0) {
vex_printf("dis_int_load(ppc)(Ox1F,b0)\n");
return False;
}
switch (opc2) {
case 0x077: // lbzux (Load B & Zero, Update Indexed, PPC32 p435)
DIP("lbzux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lwzux,rA_addr|rD_addr)\n");
return False;
}
val = loadBE(Ity_I8, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom8(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x057: // lbzx (Load B & Zero, Indexed, PPC32 p436)
DIP("lbzx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I8, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom8(ty, val, False) );
break;
case 0x177: // lhaux (Load HW Alg, Update Indexed, PPC32 p447)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lhaux,rA_addr|rD_addr)\n");
return False;
}
DIP("lhaux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, True) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x157: // lhax (Load HW Alg, Indexed, PPC32 p448)
DIP("lhax r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, True) );
break;
case 0x137: // lhzux (Load HW & Zero, Update Indexed, PPC32 p452)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lhzux,rA_addr|rD_addr)\n");
return False;
}
DIP("lhzux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x117: // lhzx (Load HW & Zero, Indexed, PPC32 p453)
DIP("lhzx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I16, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom16(ty, val, False) );
break;
case 0x037: // lwzux (Load W & Zero, Update Indexed, PPC32 p462)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lwzux,rA_addr|rD_addr)\n");
return False;
}
DIP("lwzux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I32, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom32(ty, val, False) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x017: // lwzx (Load W & Zero, Indexed, PPC32 p463)
DIP("lwzx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
val = loadBE(Ity_I32, mkexpr(EA));
putIReg( rD_addr, mkWidenFrom32(ty, val, False) );
break;
/* 64bit Loads */
case 0x035: // ldux (Load DWord, Update Indexed, PPC64 p475)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(ldux,rA_addr|rD_addr)\n");
return False;
}
DIP("ldux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
putIReg( rD_addr, loadBE(Ity_I64, mkexpr(EA)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x015: // ldx (Load DWord, Indexed, PPC64 p476)
DIP("ldx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
putIReg( rD_addr, loadBE(Ity_I64, mkexpr(EA)) );
break;
case 0x175: // lwaux (Load W Alg, Update Indexed, PPC64 p501)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(lwaux,rA_addr|rD_addr)\n");
return False;
}
DIP("lwaux r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
putIReg( rD_addr,
unop(Iop_32Sto64, loadBE(Ity_I32, mkexpr(EA))) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x155: // lwax (Load W Alg, Indexed, PPC64 p502)
DIP("lwax r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
putIReg( rD_addr,
unop(Iop_32Sto64, loadBE(Ity_I32, mkexpr(EA))) );
break;
default:
vex_printf("dis_int_load(ppc)(opc2)\n");
return False;
}
break;
/* DS Form - 64bit Loads. In each case EA will have been formed
with the lowest 2 bits masked off the immediate offset. */
case 0x3A:
switch ((b1<<1) | b0) {
case 0x0: // ld (Load DWord, PPC64 p472)
DIP("ld r%u,%d(r%u)\n", rD_addr, simm16, rA_addr);
putIReg( rD_addr, loadBE(Ity_I64, mkexpr(EA)) );
break;
case 0x1: // ldu (Load DWord, Update, PPC64 p474)
if (rA_addr == 0 || rA_addr == rD_addr) {
vex_printf("dis_int_load(ppc)(ldu,rA_addr|rD_addr)\n");
return False;
}
DIP("ldu r%u,%d(r%u)\n", rD_addr, simm16, rA_addr);
putIReg( rD_addr, loadBE(Ity_I64, mkexpr(EA)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x2: // lwa (Load Word Alg, PPC64 p499)
DIP("lwa r%u,%d(r%u)\n", rD_addr, simm16, rA_addr);
putIReg( rD_addr,
unop(Iop_32Sto64, loadBE(Ity_I32, mkexpr(EA))) );
break;
default:
vex_printf("dis_int_load(ppc)(0x3A, opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_load(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Integer Store Instructions
*/
static Bool dis_int_store ( UInt theInstr, VexAbiInfo* vbi )
{
/* D-Form, X-Form, DS-Form */
UChar opc1 = ifieldOPC(theInstr);
UInt rS_addr = ifieldRegDS(theInstr);
UInt rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
UInt rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b1 = ifieldBIT1(theInstr);
UChar b0 = ifieldBIT0(theInstr);
Int simm16 = extend_s_16to32(uimm16);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rS = newTemp(ty);
IRTemp rB = newTemp(ty);
IRTemp EA = newTemp(ty);
assign( rB, getIReg(rB_addr) );
assign( rS, getIReg(rS_addr) );
switch (opc1) {
case 0x1F: // register offset
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
break;
case 0x3E: // immediate offset: 64bit: std/stdu: mask off
// lowest 2 bits of immediate before forming EA
simm16 = simm16 & 0xFFFFFFFC;
default: // immediate offset
assign( EA, ea_rAor0_simm( rA_addr, simm16 ) );
break;
}
switch (opc1) {
case 0x26: // stb (Store B, PPC32 p509)
DIP("stb r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
storeBE( mkexpr(EA), mkNarrowTo8(ty, mkexpr(rS)) );
break;
case 0x27: // stbu (Store B, Update, PPC32 p510)
if (rA_addr == 0 ) {
vex_printf("dis_int_store(ppc)(stbu,rA_addr)\n");
return False;
}
DIP("stbu r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo8(ty, mkexpr(rS)) );
break;
case 0x2C: // sth (Store HW, PPC32 p522)
DIP("sth r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
storeBE( mkexpr(EA), mkNarrowTo16(ty, mkexpr(rS)) );
break;
case 0x2D: // sthu (Store HW, Update, PPC32 p524)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(sthu,rA_addr)\n");
return False;
}
DIP("sthu r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo16(ty, mkexpr(rS)) );
break;
case 0x24: // stw (Store W, PPC32 p530)
DIP("stw r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
storeBE( mkexpr(EA), mkNarrowTo32(ty, mkexpr(rS)) );
break;
case 0x25: // stwu (Store W, Update, PPC32 p534)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(stwu,rA_addr)\n");
return False;
}
DIP("stwu r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo32(ty, mkexpr(rS)) );
break;
/* X Form : all these use EA_indexed */
case 0x1F:
if (b0 != 0) {
vex_printf("dis_int_store(ppc)(0x1F,b0)\n");
return False;
}
switch (opc2) {
case 0x0F7: // stbux (Store B, Update Indexed, PPC32 p511)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(stbux,rA_addr)\n");
return False;
}
DIP("stbux r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo8(ty, mkexpr(rS)) );
break;
case 0x0D7: // stbx (Store B Indexed, PPC32 p512)
DIP("stbx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
storeBE( mkexpr(EA), mkNarrowTo8(ty, mkexpr(rS)) );
break;
case 0x1B7: // sthux (Store HW, Update Indexed, PPC32 p525)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(sthux,rA_addr)\n");
return False;
}
DIP("sthux r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo16(ty, mkexpr(rS)) );
break;
case 0x197: // sthx (Store HW Indexed, PPC32 p526)
DIP("sthx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
storeBE( mkexpr(EA), mkNarrowTo16(ty, mkexpr(rS)) );
break;
case 0x0B7: // stwux (Store W, Update Indexed, PPC32 p535)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(stwux,rA_addr)\n");
return False;
}
DIP("stwux r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkNarrowTo32(ty, mkexpr(rS)) );
break;
case 0x097: // stwx (Store W Indexed, PPC32 p536)
DIP("stwx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
storeBE( mkexpr(EA), mkNarrowTo32(ty, mkexpr(rS)) );
break;
/* 64bit Stores */
case 0x0B5: // stdux (Store DWord, Update Indexed, PPC64 p584)
if (rA_addr == 0) {
vex_printf("dis_int_store(ppc)(stdux,rA_addr)\n");
return False;
}
DIP("stdux r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkexpr(rS) );
break;
case 0x095: // stdx (Store DWord Indexed, PPC64 p585)
DIP("stdx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
storeBE( mkexpr(EA), mkexpr(rS) );
break;
default:
vex_printf("dis_int_store(ppc)(opc2)\n");
return False;
}
break;
/* DS Form - 64bit Stores. In each case EA will have been formed
with the lowest 2 bits masked off the immediate offset. */
case 0x3E:
switch ((b1<<1) | b0) {
case 0x0: // std (Store DWord, PPC64 p580)
DIP("std r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
storeBE( mkexpr(EA), mkexpr(rS) );
break;
case 0x1: // stdu (Store DWord, Update, PPC64 p583)
DIP("stdu r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
putIReg( rA_addr, mkexpr(EA) );
storeBE( mkexpr(EA), mkexpr(rS) );
break;
default:
vex_printf("dis_int_load(ppc)(0x3A, opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_store(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Integer Load/Store Multiple Instructions
*/
static Bool dis_int_ldst_mult ( UInt theInstr )
{
/* D-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rS_addr = rD_addr;
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
Int simm16 = extend_s_16to32(uimm16);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
UInt r = 0;
UInt ea_off = 0;
IRExpr* irx_addr;
assign( EA, ea_rAor0_simm( rA_addr, simm16 ) );
switch (opc1) {
case 0x2E: // lmw (Load Multiple Word, PPC32 p454)
if (rA_addr >= rD_addr) {
vex_printf("dis_int_ldst_mult(ppc)(lmw,rA_addr)\n");
return False;
}
DIP("lmw r%u,%d(r%u)\n", rD_addr, simm16, rA_addr);
for (r = rD_addr; r <= 31; r++) {
irx_addr = binop(Iop_Add32, mkexpr(EA), mkU32(ea_off));
putIReg( r, mkWidenFrom32(ty, loadBE(Ity_I32, irx_addr ),
False) );
ea_off += 4;
}
break;
case 0x2F: // stmw (Store Multiple Word, PPC32 p527)
DIP("stmw r%u,%d(r%u)\n", rS_addr, simm16, rA_addr);
for (r = rS_addr; r <= 31; r++) {
irx_addr = binop(Iop_Add32, mkexpr(EA), mkU32(ea_off));
storeBE( irx_addr, mkNarrowTo32(ty, getIReg(r)) );
ea_off += 4;
}
break;
default:
vex_printf("dis_int_ldst_mult(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Integer Load/Store String Instructions
*/
static
void generate_lsw_sequence ( IRTemp tNBytes, // # bytes, :: Ity_I32
IRTemp EA, // EA
Int rD, // first dst register
Int maxBytes ) // 32 or 128
{
Int i, shift = 24;
IRExpr* e_nbytes = mkexpr(tNBytes);
IRExpr* e_EA = mkexpr(EA);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rD >= 0 && rD < 32);
rD--; if (rD < 0) rD = 31;
for (i = 0; i < maxBytes; i++) {
/* if (nBytes < (i+1)) goto NIA; */
stmt( IRStmt_Exit( binop(Iop_CmpLT32U, e_nbytes, mkU32(i+1)),
Ijk_Boring,
mkSzConst( ty, nextInsnAddr()), OFFB_CIA ));
/* when crossing into a new dest register, set it to zero. */
if ((i % 4) == 0) {
rD++; if (rD == 32) rD = 0;
putIReg(rD, mkSzImm(ty, 0));
shift = 24;
}
/* rD |= (8Uto32(*(EA+i))) << shift */
vassert(shift == 0 || shift == 8 || shift == 16 || shift == 24);
putIReg(
rD,
mkWidenFrom32(
ty,
binop(
Iop_Or32,
mkNarrowTo32(ty, getIReg(rD)),
binop(
Iop_Shl32,
unop(
Iop_8Uto32,
loadBE(Ity_I8,
binop(mkSzOp(ty,Iop_Add8), e_EA, mkSzImm(ty,i)))
),
mkU8(toUChar(shift))
)
),
/*Signed*/False
)
);
shift -= 8;
}
}
static
void generate_stsw_sequence ( IRTemp tNBytes, // # bytes, :: Ity_I32
IRTemp EA, // EA
Int rS, // first src register
Int maxBytes ) // 32 or 128
{
Int i, shift = 24;
IRExpr* e_nbytes = mkexpr(tNBytes);
IRExpr* e_EA = mkexpr(EA);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
vassert(rS >= 0 && rS < 32);
rS--; if (rS < 0) rS = 31;
for (i = 0; i < maxBytes; i++) {
/* if (nBytes < (i+1)) goto NIA; */
stmt( IRStmt_Exit( binop(Iop_CmpLT32U, e_nbytes, mkU32(i+1)),
Ijk_Boring,
mkSzConst( ty, nextInsnAddr() ), OFFB_CIA ));
/* check for crossing into a new src register. */
if ((i % 4) == 0) {
rS++; if (rS == 32) rS = 0;
shift = 24;
}
/* *(EA+i) = 32to8(rS >> shift) */
vassert(shift == 0 || shift == 8 || shift == 16 || shift == 24);
storeBE(
binop(mkSzOp(ty,Iop_Add8), e_EA, mkSzImm(ty,i)),
unop(Iop_32to8,
binop(Iop_Shr32,
mkNarrowTo32(ty, getIReg(rS)),
mkU8(toUChar(shift))))
);
shift -= 8;
}
}
static Bool dis_int_ldst_str ( UInt theInstr, /*OUT*/Bool* stopHere )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rS_addr = rD_addr;
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UChar NumBytes = rB_addr;
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp t_EA = newTemp(ty);
IRTemp t_nbytes = IRTemp_INVALID;
*stopHere = False;
if (opc1 != 0x1F || b0 != 0) {
vex_printf("dis_int_ldst_str(ppc)(opc1)\n");
return False;
}
switch (opc2) {
case 0x255: // lswi (Load String Word Immediate, PPC32 p455)
/* NB: does not reject the case where RA is in the range of
registers to be loaded. It should. */
DIP("lswi r%u,r%u,%d\n", rD_addr, rA_addr, NumBytes);
assign( t_EA, ea_rAor0(rA_addr) );
if (NumBytes == 8 && !mode64) {
/* Special case hack */
/* rD = Mem[EA]; (rD+1)%32 = Mem[EA+4] */
putIReg( rD_addr,
loadBE(Ity_I32, mkexpr(t_EA)) );
putIReg( (rD_addr+1) % 32,
loadBE(Ity_I32,
binop(Iop_Add32, mkexpr(t_EA), mkU32(4))) );
} else {
t_nbytes = newTemp(Ity_I32);
assign( t_nbytes, mkU32(NumBytes==0 ? 32 : NumBytes) );
generate_lsw_sequence( t_nbytes, t_EA, rD_addr, 32 );
*stopHere = True;
}
return True;
case 0x215: // lswx (Load String Word Indexed, PPC32 p456)
/* NB: does not reject the case where RA is in the range of
registers to be loaded. It should. Although considering
that that can only be detected at run time, it's not easy to
do so. */
if (rD_addr == rA_addr || rD_addr == rB_addr)
return False;
if (rD_addr == 0 && rA_addr == 0)
return False;
DIP("lswx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
t_nbytes = newTemp(Ity_I32);
assign( t_EA, ea_rAor0_idxd(rA_addr,rB_addr) );
assign( t_nbytes, unop( Iop_8Uto32, getXER_BC() ) );
generate_lsw_sequence( t_nbytes, t_EA, rD_addr, 128 );
*stopHere = True;
return True;
case 0x2D5: // stswi (Store String Word Immediate, PPC32 p528)
DIP("stswi r%u,r%u,%d\n", rS_addr, rA_addr, NumBytes);
assign( t_EA, ea_rAor0(rA_addr) );
if (NumBytes == 8 && !mode64) {
/* Special case hack */
/* Mem[EA] = rD; Mem[EA+4] = (rD+1)%32 */
storeBE( mkexpr(t_EA),
getIReg(rD_addr) );
storeBE( binop(Iop_Add32, mkexpr(t_EA), mkU32(4)),
getIReg((rD_addr+1) % 32) );
} else {
t_nbytes = newTemp(Ity_I32);
assign( t_nbytes, mkU32(NumBytes==0 ? 32 : NumBytes) );
generate_stsw_sequence( t_nbytes, t_EA, rD_addr, 32 );
*stopHere = True;
}
return True;
case 0x295: // stswx (Store String Word Indexed, PPC32 p529)
DIP("stswx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
t_nbytes = newTemp(Ity_I32);
assign( t_EA, ea_rAor0_idxd(rA_addr,rB_addr) );
assign( t_nbytes, unop( Iop_8Uto32, getXER_BC() ) );
generate_stsw_sequence( t_nbytes, t_EA, rS_addr, 128 );
*stopHere = True;
return True;
default:
vex_printf("dis_int_ldst_str(ppc)(opc2)\n");
return False;
}
return True;
}
/* ------------------------------------------------------------------
Integer Branch Instructions
------------------------------------------------------------------ */
/*
Branch helper function
ok = BO[2] | ((CTR[0] != 0) ^ BO[1])
Returns an I32 which is 0x00000000 if the ctr condition failed
and 0xFFFFFFFF otherwise.
*/
static IRExpr* /* :: Ity_I32 */ branch_ctr_ok( UInt BO )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp ok = newTemp(Ity_I32);
if ((BO >> 2) & 1) { // independent of ctr
assign( ok, mkU32(0xFFFFFFFF) );
} else {
if ((BO >> 1) & 1) { // ctr == 0 ?
assign( ok, unop( Iop_1Sto32,
binop( mkSzOp(ty, Iop_CmpEQ8),
getGST( PPC_GST_CTR ),
mkSzImm(ty,0))) );
} else { // ctr != 0 ?
assign( ok, unop( Iop_1Sto32,
binop( mkSzOp(ty, Iop_CmpNE8),
getGST( PPC_GST_CTR ),
mkSzImm(ty,0))) );
}
}
return mkexpr(ok);
}
/*
Branch helper function cond_ok = BO[4] | (CR[BI] == BO[3])
Returns an I32 which is either 0 if the condition failed or
some arbitrary nonzero value otherwise. */
static IRExpr* /* :: Ity_I32 */ branch_cond_ok( UInt BO, UInt BI )
{
Int where;
IRTemp res = newTemp(Ity_I32);
IRTemp cr_bi = newTemp(Ity_I32);
if ((BO >> 4) & 1) {
assign( res, mkU32(1) );
} else {
// ok = (CR[BI] == BO[3]) Note, the following relies on
// getCRbit_anywhere returning a value which
// is either zero or has exactly 1 bit set.
assign( cr_bi, getCRbit_anywhere( BI, &where ) );
if ((BO >> 3) & 1) {
/* We can use cr_bi as-is. */
assign( res, mkexpr(cr_bi) );
} else {
/* We have to invert the sense of the information held in
cr_bi. For that we need to know which bit
getCRbit_anywhere regards as significant. */
assign( res, binop(Iop_Xor32, mkexpr(cr_bi),
mkU32(1<<where)) );
}
}
return mkexpr(res);
}
/*
Integer Branch Instructions
*/
static Bool dis_branch ( UInt theInstr,
VexAbiInfo* vbi,
/*OUT*/DisResult* dres,
Bool (*resteerOkFn)(void*,Addr64),
void* callback_opaque )
{
UChar opc1 = ifieldOPC(theInstr);
UChar BO = ifieldRegDS(theInstr);
UChar BI = ifieldRegA(theInstr);
UInt BD_u16 = ifieldUIMM16(theInstr) & 0xFFFFFFFC; /* mask off */
UChar b11to15 = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UInt LI_u26 = ifieldUIMM26(theInstr) & 0xFFFFFFFC; /* mask off */
UChar flag_AA = ifieldBIT1(theInstr);
UChar flag_LK = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
Addr64 tgt = 0;
Int BD = extend_s_16to32(BD_u16);
IRTemp do_branch = newTemp(Ity_I32);
IRTemp ctr_ok = newTemp(Ity_I32);
IRTemp cond_ok = newTemp(Ity_I32);
IRExpr* e_nia = mkSzImm(ty, nextInsnAddr());
IRConst* c_nia = mkSzConst(ty, nextInsnAddr());
IRTemp lr_old = newTemp(ty);
/* Hack to pass through code that just wants to read the PC */
if (theInstr == 0x429F0005) {
DIP("bcl 0x%x, 0x%x (a.k.a mr lr,cia+4)\n", BO, BI);
putGST( PPC_GST_LR, e_nia );
return True;
}
/* The default what-next. Individual cases can override it. */
dres->whatNext = Dis_StopHere;
vassert(dres->jk_StopHere == Ijk_INVALID);
switch (opc1) {
case 0x12: // b (Branch, PPC32 p360)
if (flag_AA) {
tgt = mkSzAddr( ty, extend_s_26to64(LI_u26) );
} else {
tgt = mkSzAddr( ty, guest_CIA_curr_instr +
(Long)extend_s_26to64(LI_u26) );
}
if (mode64) {
DIP("b%s%s 0x%llx\n",
flag_LK ? "l" : "", flag_AA ? "a" : "", tgt);
} else {
DIP("b%s%s 0x%x\n",
flag_LK ? "l" : "", flag_AA ? "a" : "", (Addr32)tgt);
}
if (flag_LK) {
putGST( PPC_GST_LR, e_nia );
if (vbi->guest_ppc_zap_RZ_at_bl
&& vbi->guest_ppc_zap_RZ_at_bl( (ULong)tgt) ) {
IRTemp t_tgt = newTemp(ty);
assign(t_tgt, mode64 ? mkU64(tgt) : mkU32(tgt) );
make_redzone_AbiHint( vbi, t_tgt,
"branch-and-link (unconditional call)" );
}
}
if (resteerOkFn( callback_opaque, tgt )) {
dres->whatNext = Dis_ResteerU;
dres->continueAt = tgt;
} else {
dres->jk_StopHere = flag_LK ? Ijk_Call : Ijk_Boring; ;
putGST( PPC_GST_CIA, mkSzImm(ty, tgt) );
}
break;
case 0x10: // bc (Branch Conditional, PPC32 p361)
DIP("bc%s%s 0x%x, 0x%x, 0x%x\n",
flag_LK ? "l" : "", flag_AA ? "a" : "", BO, BI, BD);
if (!(BO & 0x4)) {
putGST( PPC_GST_CTR,
binop(mkSzOp(ty, Iop_Sub8),
getGST( PPC_GST_CTR ), mkSzImm(ty, 1)) );
}
/* This is a bit subtle. ctr_ok is either all 0s or all 1s.
cond_ok is either zero or nonzero, since that's the cheapest
way to compute it. Anding them together gives a value which
is either zero or non zero and so that's what we must test
for in the IRStmt_Exit. */
assign( ctr_ok, branch_ctr_ok( BO ) );
assign( cond_ok, branch_cond_ok( BO, BI ) );
assign( do_branch,
binop(Iop_And32, mkexpr(cond_ok), mkexpr(ctr_ok)) );
if (flag_AA) {
tgt = mkSzAddr(ty, extend_s_16to64(BD_u16));
} else {
tgt = mkSzAddr(ty, guest_CIA_curr_instr +
(Long)extend_s_16to64(BD_u16));
}
if (flag_LK)
putGST( PPC_GST_LR, e_nia );
stmt( IRStmt_Exit(
binop(Iop_CmpNE32, mkexpr(do_branch), mkU32(0)),
flag_LK ? Ijk_Call : Ijk_Boring,
mkSzConst(ty, tgt), OFFB_CIA ) );
dres->jk_StopHere = Ijk_Boring;
putGST( PPC_GST_CIA, e_nia );
break;
case 0x13:
/* For bclr and bcctr, it appears that the lowest two bits of
b11to15 are a branch hint, and so we only need to ensure it's
of the form 000XX. */
if ((b11to15 & ~3) != 0) {
vex_printf("dis_int_branch(ppc)(0x13,b11to15)(%d)\n", (Int)b11to15);
return False;
}
switch (opc2) {
case 0x210: // bcctr (Branch Cond. to Count Register, PPC32 p363)
if ((BO & 0x4) == 0) { // "decr and test CTR" option invalid
vex_printf("dis_int_branch(ppc)(bcctr,BO)\n");
return False;
}
DIP("bcctr%s 0x%x, 0x%x\n", flag_LK ? "l" : "", BO, BI);
assign( cond_ok, branch_cond_ok( BO, BI ) );
/* FIXME: this is confusing. lr_old holds the old value
of ctr, not lr :-) */
assign( lr_old, addr_align( getGST( PPC_GST_CTR ), 4 ));
if (flag_LK)
putGST( PPC_GST_LR, e_nia );
stmt( IRStmt_Exit(
binop(Iop_CmpEQ32, mkexpr(cond_ok), mkU32(0)),
Ijk_Boring,
c_nia, OFFB_CIA ));
if (flag_LK && vbi->guest_ppc_zap_RZ_at_bl) {
make_redzone_AbiHint( vbi, lr_old,
"b-ctr-l (indirect call)" );
}
dres->jk_StopHere = flag_LK ? Ijk_Call : Ijk_Boring;;
putGST( PPC_GST_CIA, mkexpr(lr_old) );
break;
case 0x010: { // bclr (Branch Cond. to Link Register, PPC32 p365)
Bool vanilla_return = False;
if ((BO & 0x14 /* 1z1zz */) == 0x14 && flag_LK == 0) {
DIP("blr\n");
vanilla_return = True;
} else {
DIP("bclr%s 0x%x, 0x%x\n", flag_LK ? "l" : "", BO, BI);
}
if (!(BO & 0x4)) {
putGST( PPC_GST_CTR,
binop(mkSzOp(ty, Iop_Sub8),
getGST( PPC_GST_CTR ), mkSzImm(ty, 1)) );
}
/* See comments above for 'bc' about this */
assign( ctr_ok, branch_ctr_ok( BO ) );
assign( cond_ok, branch_cond_ok( BO, BI ) );
assign( do_branch,
binop(Iop_And32, mkexpr(cond_ok), mkexpr(ctr_ok)) );
assign( lr_old, addr_align( getGST( PPC_GST_LR ), 4 ));
if (flag_LK)
putGST( PPC_GST_LR, e_nia );
stmt( IRStmt_Exit(
binop(Iop_CmpEQ32, mkexpr(do_branch), mkU32(0)),
Ijk_Boring,
c_nia, OFFB_CIA ));
if (vanilla_return && vbi->guest_ppc_zap_RZ_at_blr) {
make_redzone_AbiHint( vbi, lr_old,
"branch-to-lr (unconditional return)" );
}
/* blrl is pretty strange; it's like a return that sets the
return address of its caller to the insn following this
one. Mark it as a return. */
dres->jk_StopHere = Ijk_Ret; /* was flag_LK ? Ijk_Call : Ijk_Ret; */
putGST( PPC_GST_CIA, mkexpr(lr_old) );
break;
}
default:
vex_printf("dis_int_branch(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_int_branch(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Condition Register Logical Instructions
*/
static Bool dis_cond_logic ( UInt theInstr )
{
/* XL-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar crbD_addr = ifieldRegDS(theInstr);
UChar crfD_addr = toUChar( IFIELD(theInstr, 23, 3) );
UChar crbA_addr = ifieldRegA(theInstr);
UChar crfS_addr = toUChar( IFIELD(theInstr, 18, 3) );
UChar crbB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRTemp crbD = newTemp(Ity_I32);
IRTemp crbA = newTemp(Ity_I32);
IRTemp crbB = newTemp(Ity_I32);
if (opc1 != 19 || b0 != 0) {
vex_printf("dis_cond_logic(ppc)(opc1)\n");
return False;
}
if (opc2 == 0) { // mcrf (Move Cond Reg Field, PPC32 p464)
if (((crbD_addr & 0x3) != 0) ||
((crbA_addr & 0x3) != 0) || (crbB_addr != 0)) {
vex_printf("dis_cond_logic(ppc)(crbD|crbA|crbB != 0)\n");
return False;
}
DIP("mcrf cr%u,cr%u\n", crfD_addr, crfS_addr);
putCR0( crfD_addr, getCR0( crfS_addr) );
putCR321( crfD_addr, getCR321(crfS_addr) );
} else {
assign( crbA, getCRbit(crbA_addr) );
if (crbA_addr == crbB_addr)
crbB = crbA;
else
assign( crbB, getCRbit(crbB_addr) );
switch (opc2) {
case 0x101: // crand (Cond Reg AND, PPC32 p372)
DIP("crand crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, binop(Iop_And32, mkexpr(crbA), mkexpr(crbB)) );
break;
case 0x081: // crandc (Cond Reg AND w. Complement, PPC32 p373)
DIP("crandc crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, binop(Iop_And32,
mkexpr(crbA),
unop(Iop_Not32, mkexpr(crbB))) );
break;
case 0x121: // creqv (Cond Reg Equivalent, PPC32 p374)
DIP("creqv crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, unop(Iop_Not32,
binop(Iop_Xor32, mkexpr(crbA), mkexpr(crbB))) );
break;
case 0x0E1: // crnand (Cond Reg NAND, PPC32 p375)
DIP("crnand crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, unop(Iop_Not32,
binop(Iop_And32, mkexpr(crbA), mkexpr(crbB))) );
break;
case 0x021: // crnor (Cond Reg NOR, PPC32 p376)
DIP("crnor crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, unop(Iop_Not32,
binop(Iop_Or32, mkexpr(crbA), mkexpr(crbB))) );
break;
case 0x1C1: // cror (Cond Reg OR, PPC32 p377)
DIP("cror crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, binop(Iop_Or32, mkexpr(crbA), mkexpr(crbB)) );
break;
case 0x1A1: // crorc (Cond Reg OR w. Complement, PPC32 p378)
DIP("crorc crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, binop(Iop_Or32,
mkexpr(crbA),
unop(Iop_Not32, mkexpr(crbB))) );
break;
case 0x0C1: // crxor (Cond Reg XOR, PPC32 p379)
DIP("crxor crb%d,crb%d,crb%d\n", crbD_addr, crbA_addr, crbB_addr);
assign( crbD, binop(Iop_Xor32, mkexpr(crbA), mkexpr(crbB)) );
break;
default:
vex_printf("dis_cond_logic(ppc)(opc2)\n");
return False;
}
putCRbit( crbD_addr, mkexpr(crbD) );
}
return True;
}
/*
Trap instructions
*/
/* Do the code generation for a trap. Returned Bool is true iff
this is an unconditional trap. If the two arg IRExpr*s are
Ity_I32s then the comparison is 32-bit. If they are Ity_I64s
then they are 64-bit, and we must be disassembling 64-bit
instructions. */
static Bool do_trap ( UChar TO,
IRExpr* argL0, IRExpr* argR0, Addr64 cia )
{
IRTemp argL, argR;
IRExpr *argLe, *argRe, *cond, *tmp;
Bool is32bit = typeOfIRExpr(irsb->tyenv, argL0 ) == Ity_I32;
IROp opAND = is32bit ? Iop_And32 : Iop_And64;
IROp opOR = is32bit ? Iop_Or32 : Iop_Or64;
IROp opCMPORDS = is32bit ? Iop_CmpORD32S : Iop_CmpORD64S;
IROp opCMPORDU = is32bit ? Iop_CmpORD32U : Iop_CmpORD64U;
IROp opCMPNE = is32bit ? Iop_CmpNE32 : Iop_CmpNE64;
IROp opCMPEQ = is32bit ? Iop_CmpEQ32 : Iop_CmpEQ64;
IRExpr* const0 = is32bit ? mkU32(0) : mkU64(0);
IRExpr* const2 = is32bit ? mkU32(2) : mkU64(2);
IRExpr* const4 = is32bit ? mkU32(4) : mkU64(4);
IRExpr* const8 = is32bit ? mkU32(8) : mkU64(8);
const UChar b11100 = 0x1C;
const UChar b00111 = 0x07;
if (is32bit) {
vassert( typeOfIRExpr(irsb->tyenv, argL0) == Ity_I32 );
vassert( typeOfIRExpr(irsb->tyenv, argR0) == Ity_I32 );
} else {
vassert( typeOfIRExpr(irsb->tyenv, argL0) == Ity_I64 );
vassert( typeOfIRExpr(irsb->tyenv, argR0) == Ity_I64 );
vassert( mode64 );
}
if ((TO & b11100) == b11100 || (TO & b00111) == b00111) {
/* Unconditional trap. Just do the exit without
testing the arguments. */
stmt( IRStmt_Exit(
binop(opCMPEQ, const0, const0),
Ijk_SigTRAP,
mode64 ? IRConst_U64(cia) : IRConst_U32((UInt)cia),
OFFB_CIA
));
return True; /* unconditional trap */
}
if (is32bit) {
argL = newTemp(Ity_I32);
argR = newTemp(Ity_I32);
} else {
argL = newTemp(Ity_I64);
argR = newTemp(Ity_I64);
}
assign( argL, argL0 );
assign( argR, argR0 );
argLe = mkexpr(argL);
argRe = mkexpr(argR);
cond = const0;
if (TO & 16) { // L <s R
tmp = binop(opAND, binop(opCMPORDS, argLe, argRe), const8);
cond = binop(opOR, tmp, cond);
}
if (TO & 8) { // L >s R
tmp = binop(opAND, binop(opCMPORDS, argLe, argRe), const4);
cond = binop(opOR, tmp, cond);
}
if (TO & 4) { // L == R
tmp = binop(opAND, binop(opCMPORDS, argLe, argRe), const2);
cond = binop(opOR, tmp, cond);
}
if (TO & 2) { // L <u R
tmp = binop(opAND, binop(opCMPORDU, argLe, argRe), const8);
cond = binop(opOR, tmp, cond);
}
if (TO & 1) { // L >u R
tmp = binop(opAND, binop(opCMPORDU, argLe, argRe), const4);
cond = binop(opOR, tmp, cond);
}
stmt( IRStmt_Exit(
binop(opCMPNE, cond, const0),
Ijk_SigTRAP,
mode64 ? IRConst_U64(cia) : IRConst_U32((UInt)cia),
OFFB_CIA
));
return False; /* not an unconditional trap */
}
static Bool dis_trapi ( UInt theInstr,
/*OUT*/DisResult* dres )
{
/* D-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar TO = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
ULong simm16 = extend_s_16to64(uimm16);
Addr64 cia = guest_CIA_curr_instr;
IRType ty = mode64 ? Ity_I64 : Ity_I32;
Bool uncond = False;
switch (opc1) {
case 0x03: // twi (Trap Word Immediate, PPC32 p548)
uncond = do_trap( TO,
mode64 ? unop(Iop_64to32, getIReg(rA_addr))
: getIReg(rA_addr),
mkU32( (UInt)simm16 ),
cia );
if (TO == 4) {
DIP("tweqi r%u,%d\n", (UInt)rA_addr, (Int)simm16);
} else {
DIP("tw%di r%u,%d\n", (Int)TO, (UInt)rA_addr, (Int)simm16);
}
break;
case 0x02: // tdi
if (!mode64)
return False;
uncond = do_trap( TO, getIReg(rA_addr), mkU64( (ULong)simm16 ), cia );
if (TO == 4) {
DIP("tdeqi r%u,%d\n", (UInt)rA_addr, (Int)simm16);
} else {
DIP("td%di r%u,%d\n", (Int)TO, (UInt)rA_addr, (Int)simm16);
}
break;
default:
return False;
}
if (uncond) {
/* If the trap shows signs of being unconditional, don't
continue decoding past it. */
putGST( PPC_GST_CIA, mkSzImm( ty, nextInsnAddr() ));
dres->jk_StopHere = Ijk_Boring;
dres->whatNext = Dis_StopHere;
}
return True;
}
static Bool dis_trap ( UInt theInstr,
/*OUT*/DisResult* dres )
{
/* X-Form */
UInt opc2 = ifieldOPClo10(theInstr);
UChar TO = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
Addr64 cia = guest_CIA_curr_instr;
IRType ty = mode64 ? Ity_I64 : Ity_I32;
Bool uncond = False;
if (ifieldBIT0(theInstr) != 0)
return False;
switch (opc2) {
case 0x004: // tw (Trap Word, PPC64 p540)
uncond = do_trap( TO,
mode64 ? unop(Iop_64to32, getIReg(rA_addr))
: getIReg(rA_addr),
mode64 ? unop(Iop_64to32, getIReg(rB_addr))
: getIReg(rB_addr),
cia );
if (TO == 4) {
DIP("tweq r%u,r%u\n", (UInt)rA_addr, (UInt)rB_addr);
} else {
DIP("tw%d r%u,r%u\n", (Int)TO, (UInt)rA_addr, (UInt)rB_addr);
}
break;
case 0x044: // td (Trap Doubleword, PPC64 p534)
if (!mode64)
return False;
uncond = do_trap( TO, getIReg(rA_addr), getIReg(rB_addr), cia );
if (TO == 4) {
DIP("tdeq r%u,r%u\n", (UInt)rA_addr, (UInt)rB_addr);
} else {
DIP("td%d r%u,r%u\n", (Int)TO, (UInt)rA_addr, (UInt)rB_addr);
}
break;
default:
return False;
}
if (uncond) {
/* If the trap shows signs of being unconditional, don't
continue decoding past it. */
putGST( PPC_GST_CIA, mkSzImm( ty, nextInsnAddr() ));
dres->jk_StopHere = Ijk_Boring;
dres->whatNext = Dis_StopHere;
}
return True;
}
/*
System Linkage Instructions
*/
static Bool dis_syslink ( UInt theInstr,
VexAbiInfo* abiinfo, DisResult* dres )
{
IRType ty = mode64 ? Ity_I64 : Ity_I32;
if (theInstr != 0x44000002) {
vex_printf("dis_syslink(ppc)(theInstr)\n");
return False;
}
// sc (System Call, PPC32 p504)
DIP("sc\n");
/* Copy CIA into the IP_AT_SYSCALL pseudo-register, so that on AIX
Valgrind can back the guest up to this instruction if it needs
to restart the syscall. */
putGST( PPC_GST_IP_AT_SYSCALL, getGST( PPC_GST_CIA ) );
/* It's important that all ArchRegs carry their up-to-date value
at this point. So we declare an end-of-block here, which
forces any TempRegs caching ArchRegs to be flushed. */
putGST( PPC_GST_CIA, abiinfo->guest_ppc_sc_continues_at_LR
? getGST( PPC_GST_LR )
: mkSzImm( ty, nextInsnAddr() ));
dres->whatNext = Dis_StopHere;
dres->jk_StopHere = Ijk_Sys_syscall;
return True;
}
/*
Memory Synchronization Instructions
Note on Reservations:
We rely on the assumption that V will in fact only allow one thread at
once to run. In effect, a thread can make a reservation, but we don't
check any stores it does. Instead, the reservation is cancelled when
the scheduler switches to another thread (run_thread_for_a_while()).
*/
static Bool dis_memsync ( UInt theInstr )
{
/* X-Form, XL-Form */
UChar opc1 = ifieldOPC(theInstr);
UInt b11to25 = IFIELD(theInstr, 11, 15);
UChar flag_L = ifieldRegDS(theInstr);
UInt b11to20 = IFIELD(theInstr, 11, 10);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rS_addr = rD_addr;
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
switch (opc1) {
/* XL-Form */
case 0x13: // isync (Instruction Synchronize, PPC32 p432)
if (opc2 != 0x096) {
vex_printf("dis_memsync(ppc)(0x13,opc2)\n");
return False;
}
if (b11to25 != 0 || b0 != 0) {
vex_printf("dis_memsync(ppc)(0x13,b11to25|b0)\n");
return False;
}
DIP("isync\n");
stmt( IRStmt_MBE(Imbe_Fence) );
break;
/* X-Form */
case 0x1F:
switch (opc2) {
case 0x356: // eieio (Enforce In-Order Exec of I/O, PPC32 p394)
if (b11to25 != 0 || b0 != 0) {
vex_printf("dis_memsync(ppc)(eiei0,b11to25|b0)\n");
return False;
}
DIP("eieio\n");
/* Insert a memory fence, just to be on the safe side. */
stmt( IRStmt_MBE(Imbe_Fence) );
break;
case 0x014: { // lwarx (Load Word and Reserve Indexed, PPC32 p458)
IRTemp res;
/* According to the PowerPC ISA version 2.05, b0 (called EH
in the documentation) is merely a hint bit to the
hardware, I think as to whether or not contention is
likely. So we can just ignore it. */
DIP("lwarx r%u,r%u,r%u,EH=%u\n", rD_addr, rA_addr, rB_addr, (UInt)b0);
// trap if misaligned
gen_SIGBUS_if_misaligned( EA, 4 );
// and actually do the load
res = newTemp(Ity_I32);
stmt( IRStmt_LLSC(Iend_BE, res, mkexpr(EA), NULL/*this is a load*/) );
putIReg( rD_addr, mkWidenFrom32(ty, mkexpr(res), False) );
break;
}
case 0x096: {
// stwcx. (Store Word Conditional Indexed, PPC32 p532)
// Note this has to handle stwcx. in both 32- and 64-bit modes,
// so isn't quite as straightforward as it might otherwise be.
IRTemp rS = newTemp(Ity_I32);
IRTemp resSC;
if (b0 != 1) {
vex_printf("dis_memsync(ppc)(stwcx.,b0)\n");
return False;
}
DIP("stwcx. r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
// trap if misaligned
gen_SIGBUS_if_misaligned( EA, 4 );
// Get the data to be stored, and narrow to 32 bits if necessary
assign( rS, mkNarrowTo32(ty, getIReg(rS_addr)) );
// Do the store, and get success/failure bit into resSC
resSC = newTemp(Ity_I1);
stmt( IRStmt_LLSC(Iend_BE, resSC, mkexpr(EA), mkexpr(rS)) );
// Set CR0[LT GT EQ S0] = 0b000 || XER[SO] on failure
// Set CR0[LT GT EQ S0] = 0b001 || XER[SO] on success
putCR321(0, binop(Iop_Shl8, unop(Iop_1Uto8, mkexpr(resSC)), mkU8(1)));
putCR0(0, getXER_SO());
/* Note:
If resaddr != lwarx_resaddr, CR0[EQ] is undefined, and
whether rS is stored is dependent on that value. */
/* So I guess we can just ignore this case? */
break;
}
case 0x256: // sync (Synchronize, PPC32 p543),
// also lwsync (L==1), ptesync (L==2)
/* http://sources.redhat.com/ml/binutils/2000-12/msg00311.html
The PowerPC architecture used in IBM chips has expanded
the sync instruction into two variants: lightweight sync
and heavyweight sync. The original sync instruction is
the new heavyweight sync and lightweight sync is a strict
subset of the heavyweight sync functionality. This allows
the programmer to specify a less expensive operation on
high-end systems when the full sync functionality is not
necessary.
The basic "sync" mnemonic now utilizes an operand. "sync"
without an operand now becomes a extended mnemonic for
heavyweight sync. Processors without the lwsync
instruction will not decode the L field and will perform a
heavyweight sync. Everything is backward compatible.
sync = sync 0
lwsync = sync 1
ptesync = sync 2 *** TODO - not implemented ***
*/
if (b11to20 != 0 || b0 != 0) {
vex_printf("dis_memsync(ppc)(sync/lwsync,b11to20|b0)\n");
return False;
}
if (flag_L != 0/*sync*/ && flag_L != 1/*lwsync*/) {
vex_printf("dis_memsync(ppc)(sync/lwsync,flag_L)\n");
return False;
}
DIP("%ssync\n", flag_L == 1 ? "lw" : "");
/* Insert a memory fence. It's sometimes important that these
are carried through to the generated code. */
stmt( IRStmt_MBE(Imbe_Fence) );
break;
/* 64bit Memsync */
case 0x054: { // ldarx (Load DWord and Reserve Indexed, PPC64 p473)
IRTemp res;
/* According to the PowerPC ISA version 2.05, b0 (called EH
in the documentation) is merely a hint bit to the
hardware, I think as to whether or not contention is
likely. So we can just ignore it. */
if (!mode64)
return False;
DIP("ldarx r%u,r%u,r%u,EH=%u\n", rD_addr, rA_addr, rB_addr, (UInt)b0);
// trap if misaligned
gen_SIGBUS_if_misaligned( EA, 8 );
// and actually do the load
res = newTemp(Ity_I64);
stmt( IRStmt_LLSC(Iend_BE, res, mkexpr(EA), NULL/*this is a load*/) );
putIReg( rD_addr, mkexpr(res) );
break;
}
case 0x0D6: { // stdcx. (Store DWord Condition Indexd, PPC64 p581)
// A marginally simplified version of the stwcx. case
IRTemp rS = newTemp(Ity_I64);
IRTemp resSC;
if (b0 != 1) {
vex_printf("dis_memsync(ppc)(stdcx.,b0)\n");
return False;
}
if (!mode64)
return False;
DIP("stdcx. r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
// trap if misaligned
gen_SIGBUS_if_misaligned( EA, 8 );
// Get the data to be stored
assign( rS, getIReg(rS_addr) );
// Do the store, and get success/failure bit into resSC
resSC = newTemp(Ity_I1);
stmt( IRStmt_LLSC(Iend_BE, resSC, mkexpr(EA), mkexpr(rS)) );
// Set CR0[LT GT EQ S0] = 0b000 || XER[SO] on failure
// Set CR0[LT GT EQ S0] = 0b001 || XER[SO] on success
putCR321(0, binop(Iop_Shl8, unop(Iop_1Uto8, mkexpr(resSC)), mkU8(1)));
putCR0(0, getXER_SO());
/* Note:
If resaddr != lwarx_resaddr, CR0[EQ] is undefined, and
whether rS is stored is dependent on that value. */
/* So I guess we can just ignore this case? */
break;
}
default:
vex_printf("dis_memsync(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_memsync(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Integer Shift Instructions
*/
static Bool dis_int_shift ( UInt theInstr )
{
/* X-Form, XS-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UChar sh_imm = rB_addr;
UInt opc2 = ifieldOPClo10(theInstr);
UChar b1 = ifieldBIT1(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rA = newTemp(ty);
IRTemp rS = newTemp(ty);
IRTemp rB = newTemp(ty);
IRTemp outofrange = newTemp(Ity_I8);
IRTemp rS_lo32 = newTemp(Ity_I32);
IRTemp rB_lo32 = newTemp(Ity_I32);
IRExpr* e_tmp;
assign( rS, getIReg(rS_addr) );
assign( rB, getIReg(rB_addr) );
assign( rS_lo32, mkNarrowTo32(ty, mkexpr(rS)) );
assign( rB_lo32, mkNarrowTo32(ty, mkexpr(rB)) );
if (opc1 == 0x1F) {
switch (opc2) {
case 0x018: { // slw (Shift Left Word, PPC32 p505)
DIP("slw%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr);
/* rA = rS << rB */
/* ppc32 semantics are:
slw(x,y) = (x << (y & 31)) -- primary result
& ~((y << 26) >>s 31) -- make result 0
for y in 32 .. 63
*/
e_tmp =
binop( Iop_And32,
binop( Iop_Shl32,
mkexpr(rS_lo32),
unop( Iop_32to8,
binop(Iop_And32,
mkexpr(rB_lo32), mkU32(31)))),
unop( Iop_Not32,
binop( Iop_Sar32,
binop(Iop_Shl32, mkexpr(rB_lo32), mkU8(26)),
mkU8(31))) );
assign( rA, mkWidenFrom32(ty, e_tmp, /* Signed */False) );
break;
}
case 0x318: { // sraw (Shift Right Alg Word, PPC32 p506)
IRTemp sh_amt = newTemp(Ity_I32);
DIP("sraw%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr);
/* JRS: my reading of the (poorly worded) PPC32 doc p506 is:
amt = rB & 63
rA = Sar32( rS, amt > 31 ? 31 : amt )
XER.CA = amt > 31 ? sign-of-rS : (computation as per srawi)
*/
assign( sh_amt, binop(Iop_And32, mkU32(0x3F),
mkexpr(rB_lo32)) );
assign( outofrange,
unop( Iop_1Uto8,
binop(Iop_CmpLT32U, mkU32(31),
mkexpr(sh_amt)) ));
e_tmp = binop( Iop_Sar32,
mkexpr(rS_lo32),
unop( Iop_32to8,
IRExpr_Mux0X( mkexpr(outofrange),
mkexpr(sh_amt),
mkU32(31)) ) );
assign( rA, mkWidenFrom32(ty, e_tmp, /* Signed */True) );
set_XER_CA( ty, PPCG_FLAG_OP_SRAW,
mkexpr(rA),
mkWidenFrom32(ty, mkexpr(rS_lo32), True),
mkWidenFrom32(ty, mkexpr(sh_amt), True ),
mkWidenFrom32(ty, getXER_CA32(), True) );
break;
}
case 0x338: // srawi (Shift Right Alg Word Immediate, PPC32 p507)
DIP("srawi%s r%u,r%u,%d\n", flag_rC ? ".":"",
rA_addr, rS_addr, sh_imm);
vassert(sh_imm < 32);
if (mode64) {
assign( rA, binop(Iop_Sar64,
binop(Iop_Shl64, getIReg(rS_addr),
mkU8(32)),
mkU8(32 + sh_imm)) );
} else {
assign( rA, binop(Iop_Sar32, mkexpr(rS_lo32),
mkU8(sh_imm)) );
}
set_XER_CA( ty, PPCG_FLAG_OP_SRAWI,
mkexpr(rA),
mkWidenFrom32(ty, mkexpr(rS_lo32), /* Syned */True),
mkSzImm(ty, sh_imm),
mkWidenFrom32(ty, getXER_CA32(), /* Syned */False) );
break;
case 0x218: // srw (Shift Right Word, PPC32 p508)
DIP("srw%s r%u,r%u,r%u\n", flag_rC ? ".":"",
rA_addr, rS_addr, rB_addr);
/* rA = rS >>u rB */
/* ppc32 semantics are:
srw(x,y) = (x >>u (y & 31)) -- primary result
& ~((y << 26) >>s 31) -- make result 0
for y in 32 .. 63
*/
e_tmp =
binop(
Iop_And32,
binop( Iop_Shr32,
mkexpr(rS_lo32),
unop( Iop_32to8,
binop(Iop_And32, mkexpr(rB_lo32),
mkU32(31)))),
unop( Iop_Not32,
binop( Iop_Sar32,
binop(Iop_Shl32, mkexpr(rB_lo32),
mkU8(26)),
mkU8(31))));
assign( rA, mkWidenFrom32(ty, e_tmp, /* Signed */False) );
break;
/* 64bit Shifts */
case 0x01B: // sld (Shift Left DWord, PPC64 p568)
DIP("sld%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
/* rA = rS << rB */
/* ppc64 semantics are:
slw(x,y) = (x << (y & 63)) -- primary result
& ~((y << 57) >>s 63) -- make result 0
for y in 64 ..
*/
assign( rA,
binop(
Iop_And64,
binop( Iop_Shl64,
mkexpr(rS),
unop( Iop_64to8,
binop(Iop_And64, mkexpr(rB), mkU64(63)))),
unop( Iop_Not64,
binop( Iop_Sar64,
binop(Iop_Shl64, mkexpr(rB), mkU8(57)),
mkU8(63)))) );
break;
case 0x31A: { // srad (Shift Right Alg DWord, PPC64 p570)
IRTemp sh_amt = newTemp(Ity_I64);
DIP("srad%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
/* amt = rB & 127
rA = Sar64( rS, amt > 63 ? 63 : amt )
XER.CA = amt > 63 ? sign-of-rS : (computation as per srawi)
*/
assign( sh_amt, binop(Iop_And64, mkU64(0x7F), mkexpr(rB)) );
assign( outofrange,
unop( Iop_1Uto8,
binop(Iop_CmpLT64U, mkU64(63),
mkexpr(sh_amt)) ));
assign( rA,
binop( Iop_Sar64,
mkexpr(rS),
unop( Iop_64to8,
IRExpr_Mux0X( mkexpr(outofrange),
mkexpr(sh_amt),
mkU64(63)) ))
);
set_XER_CA( ty, PPCG_FLAG_OP_SRAD,
mkexpr(rA), mkexpr(rS), mkexpr(sh_amt),
mkWidenFrom32(ty, getXER_CA32(), /* Syned */False) );
break;
}
case 0x33A: case 0x33B: // sradi (Shr Alg DWord Imm, PPC64 p571)
sh_imm |= b1<<5;
vassert(sh_imm < 64);
DIP("sradi%s r%u,r%u,%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, sh_imm);
assign( rA, binop(Iop_Sar64, getIReg(rS_addr), mkU8(sh_imm)) );
set_XER_CA( ty, PPCG_FLAG_OP_SRADI,
mkexpr(rA),
getIReg(rS_addr),
mkU64(sh_imm),
mkWidenFrom32(ty, getXER_CA32(), /* Syned */False) );
break;
case 0x21B: // srd (Shift Right DWord, PPC64 p574)
DIP("srd%s r%u,r%u,r%u\n",
flag_rC ? ".":"", rA_addr, rS_addr, rB_addr);
/* rA = rS >>u rB */
/* ppc semantics are:
srw(x,y) = (x >>u (y & 63)) -- primary result
& ~((y << 57) >>s 63) -- make result 0
for y in 64 .. 127
*/
assign( rA,
binop(
Iop_And64,
binop( Iop_Shr64,
mkexpr(rS),
unop( Iop_64to8,
binop(Iop_And64, mkexpr(rB), mkU64(63)))),
unop( Iop_Not64,
binop( Iop_Sar64,
binop(Iop_Shl64, mkexpr(rB), mkU8(57)),
mkU8(63)))) );
break;
default:
vex_printf("dis_int_shift(ppc)(opc2)\n");
return False;
}
} else {
vex_printf("dis_int_shift(ppc)(opc1)\n");
return False;
}
putIReg( rA_addr, mkexpr(rA) );
if (flag_rC) {
set_CR0( mkexpr(rA) );
}
return True;
}
/*
Integer Load/Store Reverse Instructions
*/
/* Generates code to swap the byte order in an Ity_I32. */
static IRExpr* /* :: Ity_I32 */ gen_byterev32 ( IRTemp t )
{
vassert(typeOfIRTemp(irsb->tyenv, t) == Ity_I32);
return
binop(Iop_Or32,
binop(Iop_Shl32, mkexpr(t), mkU8(24)),
binop(Iop_Or32,
binop(Iop_And32, binop(Iop_Shl32, mkexpr(t), mkU8(8)),
mkU32(0x00FF0000)),
binop(Iop_Or32,
binop(Iop_And32, binop(Iop_Shr32, mkexpr(t), mkU8(8)),
mkU32(0x0000FF00)),
binop(Iop_And32, binop(Iop_Shr32, mkexpr(t), mkU8(24)),
mkU32(0x000000FF) )
)));
}
/* Generates code to swap the byte order in the lower half of an Ity_I32,
and zeroes the upper half. */
static IRExpr* /* :: Ity_I32 */ gen_byterev16 ( IRTemp t )
{
vassert(typeOfIRTemp(irsb->tyenv, t) == Ity_I32);
return
binop(Iop_Or32,
binop(Iop_And32, binop(Iop_Shl32, mkexpr(t), mkU8(8)),
mkU32(0x0000FF00)),
binop(Iop_And32, binop(Iop_Shr32, mkexpr(t), mkU8(8)),
mkU32(0x000000FF))
);
}
static Bool dis_int_ldst_rev ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar rD_addr = ifieldRegDS(theInstr);
UChar rS_addr = rD_addr;
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRTemp w1 = newTemp(Ity_I32);
IRTemp w2 = newTemp(Ity_I32);
if (opc1 != 0x1F || b0 != 0) {
vex_printf("dis_int_ldst_rev(ppc)(opc1|b0)\n");
return False;
}
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
switch (opc2) {
case 0x316: // lhbrx (Load Halfword Byte-Reverse Indexed, PPC32 p449)
DIP("lhbrx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
assign( w1, unop(Iop_16Uto32, loadBE(Ity_I16, mkexpr(EA))) );
assign( w2, gen_byterev16(w1) );
putIReg( rD_addr, mkWidenFrom32(ty, mkexpr(w2),
/* Signed */False) );
break;
case 0x216: // lwbrx (Load Word Byte-Reverse Indexed, PPC32 p459)
DIP("lwbrx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
assign( w1, loadBE(Ity_I32, mkexpr(EA)) );
assign( w2, gen_byterev32(w1) );
putIReg( rD_addr, mkWidenFrom32(ty, mkexpr(w2),
/* Signed */False) );
break;
case 0x214: // ldbrx (Load Doubleword Byte-Reverse Indexed)
{
IRExpr * nextAddr;
IRTemp w3 = newTemp( Ity_I32 );
IRTemp w4 = newTemp( Ity_I32 );
DIP("ldbrx r%u,r%u,r%u\n", rD_addr, rA_addr, rB_addr);
assign( w1, loadBE( Ity_I32, mkexpr( EA ) ) );
assign( w2, gen_byterev32( w1 ) );
nextAddr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( 4 ) : mkU32( 4 ) );
assign( w3, loadBE( Ity_I32, nextAddr ) );
assign( w4, gen_byterev32( w3 ) );
putIReg( rD_addr, binop( Iop_32HLto64, mkexpr( w4 ), mkexpr( w2 ) ) );
break;
}
case 0x396: // sthbrx (Store Half Word Byte-Reverse Indexed, PPC32 p523)
DIP("sthbrx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
assign( w1, mkNarrowTo32(ty, getIReg(rS_addr)) );
storeBE( mkexpr(EA), unop(Iop_32to16, gen_byterev16(w1)) );
break;
case 0x296: // stwbrx (Store Word Byte-Reverse Indxd, PPC32 p531)
DIP("stwbrx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
assign( w1, mkNarrowTo32(ty, getIReg(rS_addr)) );
storeBE( mkexpr(EA), gen_byterev32(w1) );
break;
case 0x294: // stdbrx (Store Doubleword Byte-Reverse Indexed)
{
IRTemp lo = newTemp(Ity_I32);
IRTemp hi = newTemp(Ity_I32);
IRTemp rS = newTemp(Ity_I64);
assign( rS, getIReg( rS_addr ) );
DIP("stdbrx r%u,r%u,r%u\n", rS_addr, rA_addr, rB_addr);
assign(lo, unop(Iop_64HIto32, mkexpr(rS)));
assign(hi, unop(Iop_64to32, mkexpr(rS)));
storeBE( mkexpr( EA ),
binop( Iop_32HLto64, gen_byterev32( hi ), gen_byterev32( lo ) ) );
break;
}
default:
vex_printf("dis_int_ldst_rev(ppc)(opc2)\n");
return False;
}
return True;
}
/*
Processor Control Instructions
*/
static Bool dis_proc_ctl ( VexAbiInfo* vbi, UInt theInstr )
{
UChar opc1 = ifieldOPC(theInstr);
/* X-Form */
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
UChar b21to22 = toUChar( IFIELD( theInstr, 21, 2 ) );
UChar rD_addr = ifieldRegDS(theInstr);
UInt b11to20 = IFIELD( theInstr, 11, 10 );
/* XFX-Form */
UChar rS_addr = rD_addr;
UInt SPR = b11to20;
UInt TBR = b11to20;
UChar b20 = toUChar( IFIELD( theInstr, 20, 1 ) );
UInt CRM = IFIELD( theInstr, 12, 8 );
UChar b11 = toUChar( IFIELD( theInstr, 11, 1 ) );
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp rS = newTemp(ty);
assign( rS, getIReg(rS_addr) );
/* Reorder SPR field as per PPC32 p470 */
SPR = ((SPR & 0x1F) << 5) | ((SPR >> 5) & 0x1F);
/* Reorder TBR field as per PPC32 p475 */
TBR = ((TBR & 31) << 5) | ((TBR >> 5) & 31);
if (opc1 != 0x1F || b0 != 0) {
vex_printf("dis_proc_ctl(ppc)(opc1|b0)\n");
return False;
}
switch (opc2) {
/* X-Form */
case 0x200: { // mcrxr (Move to Cond Register from XER, PPC32 p466)
if (b21to22 != 0 || b11to20 != 0) {
vex_printf("dis_proc_ctl(ppc)(mcrxr,b21to22|b11to20)\n");
return False;
}
DIP("mcrxr crf%d\n", crfD);
/* Move XER[0-3] (the top 4 bits of XER) to CR[crfD] */
putGST_field( PPC_GST_CR,
getGST_field( PPC_GST_XER, 7 ),
crfD );
// Clear XER[0-3]
putXER_SO( mkU8(0) );
putXER_OV( mkU8(0) );
putXER_CA( mkU8(0) );
break;
}
case 0x013:
// b11to20==0: mfcr (Move from Cond Register, PPC32 p467)
// b20==1 & b11==0: mfocrf (Move from One CR Field)
// However it seems that the 'mfcr' behaviour is an acceptable
// implementation of mfocr (from the 2.02 arch spec)
if (b11to20 == 0) {
DIP("mfcr r%u\n", rD_addr);
putIReg( rD_addr, mkWidenFrom32(ty, getGST( PPC_GST_CR ),
/* Signed */False) );
break;
}
if (b20 == 1 && b11 == 0) {
DIP("mfocrf r%u,%u\n", rD_addr, CRM);
putIReg( rD_addr, mkWidenFrom32(ty, getGST( PPC_GST_CR ),
/* Signed */False) );
break;
}
/* not decodable */
return False;
/* XFX-Form */
case 0x153: // mfspr (Move from Special-Purpose Register, PPC32 p470)
switch (SPR) { // Choose a register...
case 0x1:
DIP("mfxer r%u\n", rD_addr);
putIReg( rD_addr, mkWidenFrom32(ty, getGST( PPC_GST_XER ),
/* Signed */False) );
break;
case 0x8:
DIP("mflr r%u\n", rD_addr);
putIReg( rD_addr, getGST( PPC_GST_LR ) );
break;
case 0x9:
DIP("mfctr r%u\n", rD_addr);
putIReg( rD_addr, getGST( PPC_GST_CTR ) );
break;
case 0x100:
DIP("mfvrsave r%u\n", rD_addr);
putIReg( rD_addr, mkWidenFrom32(ty, getGST( PPC_GST_VRSAVE ),
/* Signed */False) );
break;
case 0x103:
DIP("mfspr r%u, SPRG3(readonly)\n", rD_addr);
putIReg( rD_addr, getGST( PPC_GST_SPRG3_RO ) );
break;
/* Even a lowly PPC7400 can run the associated helper, so no
obvious need for feature testing at this point. */
case 268 /* 0x10C */:
case 269 /* 0x10D */: {
UInt arg = SPR==268 ? 0 : 1;
IRTemp val = newTemp(Ity_I32);
IRExpr** args = mkIRExprVec_1( mkU32(arg) );
IRDirty* d = unsafeIRDirty_1_N(
val,
0/*regparms*/,
"ppc32g_dirtyhelper_MFSPR_268_269",
fnptr_to_fnentry
(vbi, &ppc32g_dirtyhelper_MFSPR_268_269),
args
);
/* execute the dirty call, dumping the result in val. */
stmt( IRStmt_Dirty(d) );
putIReg( rD_addr,
mkWidenFrom32(ty, mkexpr(val), False/*unsigned*/) );
DIP("mfspr r%u,%u", rD_addr, (UInt)SPR);
break;
}
/* Again, runs natively on PPC7400 (7447, really). Not
bothering with a feature test. */
case 287: /* 0x11F */ {
IRTemp val = newTemp(Ity_I32);
IRExpr** args = mkIRExprVec_0();
IRDirty* d = unsafeIRDirty_1_N(
val,
0/*regparms*/,
"ppc32g_dirtyhelper_MFSPR_287",
fnptr_to_fnentry
(vbi, &ppc32g_dirtyhelper_MFSPR_287),
args
);
/* execute the dirty call, dumping the result in val. */
stmt( IRStmt_Dirty(d) );
putIReg( rD_addr,
mkWidenFrom32(ty, mkexpr(val), False/*unsigned*/) );
DIP("mfspr r%u,%u", rD_addr, (UInt)SPR);
break;
}
default:
vex_printf("dis_proc_ctl(ppc)(mfspr,SPR)(0x%x)\n", SPR);
return False;
}
break;
case 0x173: { // mftb (Move from Time Base, PPC32 p475)
IRTemp val = newTemp(Ity_I64);
IRExpr** args = mkIRExprVec_0();
IRDirty* d = unsafeIRDirty_1_N(
val,
0/*regparms*/,
"ppcg_dirtyhelper_MFTB",
fnptr_to_fnentry(vbi, &ppcg_dirtyhelper_MFTB),
args );
/* execute the dirty call, dumping the result in val. */
stmt( IRStmt_Dirty(d) );
switch (TBR) {
case 269:
DIP("mftbu r%u", rD_addr);
putIReg( rD_addr,
mkWidenFrom32(ty, unop(Iop_64HIto32, mkexpr(val)),
/* Signed */False) );
break;
case 268:
DIP("mftb r%u", rD_addr);
putIReg( rD_addr, (mode64) ? mkexpr(val) :
unop(Iop_64to32, mkexpr(val)) );
break;
default:
return False; /* illegal instruction */
}
break;
}
case 0x090: {
// b20==0: mtcrf (Move to Cond Register Fields, PPC32 p477)
// b20==1: mtocrf (Move to One Cond Reg Field)
Int cr;
UChar shft;
if (b11 != 0)
return False;
if (b20 == 1) {
/* ppc64 v2.02 spec says mtocrf gives undefined outcome if >
1 field is written. It seems more robust to decline to
decode the insn if so. */
switch (CRM) {
case 0x01: case 0x02: case 0x04: case 0x08:
case 0x10: case 0x20: case 0x40: case 0x80:
break;
default:
return False;
}
}
DIP("%s 0x%x,r%u\n", b20==1 ? "mtocrf" : "mtcrf",
CRM, rS_addr);
/* Write to each field specified by CRM */
for (cr = 0; cr < 8; cr++) {
if ((CRM & (1 << (7-cr))) == 0)
continue;
shft = 4*(7-cr);
putGST_field( PPC_GST_CR,
binop(Iop_Shr32,
mkNarrowTo32(ty, mkexpr(rS)),
mkU8(shft)), cr );
}
break;
}
case 0x1D3: // mtspr (Move to Special-Purpose Register, PPC32 p483)
switch (SPR) { // Choose a register...
case 0x1:
DIP("mtxer r%u\n", rS_addr);
putGST( PPC_GST_XER, mkNarrowTo32(ty, mkexpr(rS)) );
break;
case 0x8:
DIP("mtlr r%u\n", rS_addr);
putGST( PPC_GST_LR, mkexpr(rS) );
break;
case 0x9:
DIP("mtctr r%u\n", rS_addr);
putGST( PPC_GST_CTR, mkexpr(rS) );
break;
case 0x100:
DIP("mtvrsave r%u\n", rS_addr);
putGST( PPC_GST_VRSAVE, mkNarrowTo32(ty, mkexpr(rS)) );
break;
default:
vex_printf("dis_proc_ctl(ppc)(mtspr,SPR)(%u)\n", SPR);
return False;
}
break;
default:
vex_printf("dis_proc_ctl(ppc)(opc2)\n");
return False;
}
return True;
}
/*
Cache Management Instructions
*/
static Bool dis_cache_manage ( UInt theInstr,
DisResult* dres,
VexArchInfo* guest_archinfo )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar b21to25 = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
UInt lineszB = guest_archinfo->ppc_cache_line_szB;
Bool is_dcbzl = False;
IRType ty = mode64 ? Ity_I64 : Ity_I32;
/* For dcbt, the lowest two bits of b21to25 encode an
access-direction hint (TH field) which we ignore. Well, that's
what the PowerPC documentation says. In fact xlc -O4 on POWER5
seems to generate values of 8 and 10 for b21to25. */
if (opc1 == 0x1F && opc2 == 0x116) {
/* b21to25 &= ~3; */ /* if the docs were true */
b21to25 = 0; /* blunt instrument */
}
if (opc1 == 0x1F && opc2 == 0x3F6) { // dcbz
if (b21to25 == 1) {
is_dcbzl = True;
b21to25 = 0;
if (!(guest_archinfo->ppc_dcbzl_szB)) {
vex_printf("dis_cache_manage(ppc)(dcbzl not supported by host)\n");
return False;
}
}
}
if (opc1 != 0x1F || b21to25 != 0 || b0 != 0) {
if (0) vex_printf("dis_cache_manage %d %d %d\n",
(Int)opc1, (Int)b21to25, (Int)b0);
vex_printf("dis_cache_manage(ppc)(opc1|b21to25|b0)\n");
return False;
}
/* stay sane .. */
vassert(lineszB == 32 || lineszB == 64 || lineszB == 128);
switch (opc2) {
//zz case 0x2F6: // dcba (Data Cache Block Allocate, PPC32 p380)
//zz vassert(0); /* AWAITING TEST CASE */
//zz DIP("dcba r%u,r%u\n", rA_addr, rB_addr);
//zz if (0) vex_printf("vex ppc->IR: kludged dcba\n");
//zz break;
case 0x056: // dcbf (Data Cache Block Flush, PPC32 p382)
DIP("dcbf r%u,r%u\n", rA_addr, rB_addr);
/* nop as far as vex is concerned */
break;
case 0x036: // dcbst (Data Cache Block Store, PPC32 p384)
DIP("dcbst r%u,r%u\n", rA_addr, rB_addr);
/* nop as far as vex is concerned */
break;
case 0x116: // dcbt (Data Cache Block Touch, PPC32 p385)
DIP("dcbt r%u,r%u\n", rA_addr, rB_addr);
/* nop as far as vex is concerned */
break;
case 0x0F6: // dcbtst (Data Cache Block Touch for Store, PPC32 p386)
DIP("dcbtst r%u,r%u\n", rA_addr, rB_addr);
/* nop as far as vex is concerned */
break;
case 0x3F6: { // dcbz (Data Cache Block Clear to Zero, PPC32 p387)
// dcbzl (Data Cache Block Clear to Zero Long, bug#135264)
/* Clear all bytes in cache block at (rA|0) + rB. */
IRTemp EA = newTemp(ty);
IRTemp addr = newTemp(ty);
IRExpr* irx_addr;
UInt i;
UInt clearszB;
if (is_dcbzl) {
clearszB = guest_archinfo->ppc_dcbzl_szB;
DIP("dcbzl r%u,r%u\n", rA_addr, rB_addr);
}
else {
clearszB = guest_archinfo->ppc_dcbz_szB;
DIP("dcbz r%u,r%u\n", rA_addr, rB_addr);
}
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
if (mode64) {
/* Round EA down to the start of the containing block. */
assign( addr, binop( Iop_And64,
mkexpr(EA),
mkU64( ~((ULong)clearszB-1) )) );
for (i = 0; i < clearszB / 8; i++) {
irx_addr = binop( Iop_Add64, mkexpr(addr), mkU64(i*8) );
storeBE( irx_addr, mkU64(0) );
}
} else {
/* Round EA down to the start of the containing block. */
assign( addr, binop( Iop_And32,
mkexpr(EA),
mkU32( ~(clearszB-1) )) );
for (i = 0; i < clearszB / 4; i++) {
irx_addr = binop( Iop_Add32, mkexpr(addr), mkU32(i*4) );
storeBE( irx_addr, mkU32(0) );
}
}
break;
}
case 0x3D6: {
// icbi (Instruction Cache Block Invalidate, PPC32 p431)
/* Invalidate all translations containing code from the cache
block at (rA|0) + rB. */
IRTemp EA = newTemp(ty);
IRTemp addr = newTemp(ty);
DIP("icbi r%u,r%u\n", rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
/* Round EA down to the start of the containing block. */
assign( addr, binop( mkSzOp(ty, Iop_And8),
mkexpr(EA),
mkSzImm(ty, ~(((ULong)lineszB)-1) )) );
putGST( PPC_GST_TISTART, mkexpr(addr) );
putGST( PPC_GST_TILEN, mkSzImm(ty, lineszB) );
/* be paranoid ... */
stmt( IRStmt_MBE(Imbe_Fence) );
putGST( PPC_GST_CIA, mkSzImm(ty, nextInsnAddr()));
dres->jk_StopHere = Ijk_TInval;
dres->whatNext = Dis_StopHere;
break;
}
default:
vex_printf("dis_cache_manage(ppc)(opc2)\n");
return False;
}
return True;
}
/*------------------------------------------------------------*/
/*--- Floating Point Helpers ---*/
/*------------------------------------------------------------*/
/* --------- Synthesise a 2-bit FPU rounding mode. --------- */
/* Produces a value in 0 .. 3, which is encoded as per the type
IRRoundingMode. PPCRoundingMode encoding is different to
IRRoundingMode, so need to map it.
*/
static IRExpr* /* :: Ity_I32 */ get_IR_roundingmode ( void )
{
/*
rounding mode | PPC | IR
------------------------
to nearest | 00 | 00
to zero | 01 | 11
to +infinity | 10 | 10
to -infinity | 11 | 01
*/
IRTemp rm_PPC32 = newTemp(Ity_I32);
assign( rm_PPC32, getGST_masked( PPC_GST_FPSCR, MASK_FPSCR_RN ) );
// rm_IR = XOR( rm_PPC32, (rm_PPC32 << 1) & 2)
return binop( Iop_Xor32,
mkexpr(rm_PPC32),
binop( Iop_And32,
binop(Iop_Shl32, mkexpr(rm_PPC32), mkU8(1)),
mkU32(2) ));
}
/* The DFP IR rounding modes were chosen such that the existing PPC to IR
* mapping would still work with the extended three bit DFP rounding
* mode designator.
* rounding mode | PPC | IR
* -----------------------------------------------
* to nearest, ties to even | 000 | 000
* to zero | 001 | 011
* to +infinity | 010 | 010
* to -infinity | 011 | 001
* to nearest, ties away from 0 | 100 | 100
* to nearest, ties toward 0 | 101 | 111
* to away from 0 | 110 | 110
* to prepare for shorter precision | 111 | 101
*/
static IRExpr* /* :: Ity_I32 */ get_IR_roundingmode_DFP( void )
{
IRTemp rm_PPC32 = newTemp( Ity_I32 );
assign( rm_PPC32, getGST_masked_upper( PPC_GST_FPSCR, MASK_FPSCR_DRN ) );
// rm_IR = XOR( rm_PPC32, (rm_PPC32 << 1) & 2)
return binop( Iop_Xor32,
mkexpr( rm_PPC32 ),
binop( Iop_And32,
binop( Iop_Shl32, mkexpr( rm_PPC32 ), mkU8( 1 ) ),
mkU32( 2 ) ) );
}
/*------------------------------------------------------------*/
/*--- Floating Point Instruction Translation ---*/
/*------------------------------------------------------------*/
/*
Floating Point Load Instructions
*/
static Bool dis_fp_load ( UInt theInstr )
{
/* X-Form, D-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frD_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
Int simm16 = extend_s_16to32(uimm16);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRTemp rA = newTemp(ty);
IRTemp rB = newTemp(ty);
IRTemp iHi = newTemp(Ity_I32);
IRTemp iLo = newTemp(Ity_I32);
assign( rA, getIReg(rA_addr) );
assign( rB, getIReg(rB_addr) );
/* These are completely straightforward from a rounding and status
bits perspective: no rounding involved and no funny status or CR
bits affected. */
switch (opc1) {
case 0x30: // lfs (Load Float Single, PPC32 p441)
DIP("lfs fr%u,%d(r%u)\n", frD_addr, simm16, rA_addr);
assign( EA, ea_rAor0_simm(rA_addr, simm16) );
putFReg( frD_addr,
unop(Iop_F32toF64, loadBE(Ity_F32, mkexpr(EA))) );
break;
case 0x31: // lfsu (Load Float Single, Update, PPC32 p442)
if (rA_addr == 0)
return False;
DIP("lfsu fr%u,%d(r%u)\n", frD_addr, simm16, rA_addr);
assign( EA, ea_rA_simm(rA_addr, simm16) );
putFReg( frD_addr,
unop(Iop_F32toF64, loadBE(Ity_F32, mkexpr(EA))) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x32: // lfd (Load Float Double, PPC32 p437)
DIP("lfd fr%u,%d(r%u)\n", frD_addr, simm16, rA_addr);
assign( EA, ea_rAor0_simm(rA_addr, simm16) );
putFReg( frD_addr, loadBE(Ity_F64, mkexpr(EA)) );
break;
case 0x33: // lfdu (Load Float Double, Update, PPC32 p438)
if (rA_addr == 0)
return False;
DIP("lfdu fr%u,%d(r%u)\n", frD_addr, simm16, rA_addr);
assign( EA, ea_rA_simm(rA_addr, simm16) );
putFReg( frD_addr, loadBE(Ity_F64, mkexpr(EA)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x1F:
if (b0 != 0) {
vex_printf("dis_fp_load(ppc)(instr,b0)\n");
return False;
}
switch(opc2) {
case 0x217: // lfsx (Load Float Single Indexed, PPC32 p444)
DIP("lfsx fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
putFReg( frD_addr, unop( Iop_F32toF64,
loadBE(Ity_F32, mkexpr(EA))) );
break;
case 0x237: // lfsux (Load Float Single, Update Indxd, PPC32 p443)
if (rA_addr == 0)
return False;
DIP("lfsux fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rA_idxd(rA_addr, rB_addr) );
putFReg( frD_addr,
unop(Iop_F32toF64, loadBE(Ity_F32, mkexpr(EA))) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x257: // lfdx (Load Float Double Indexed, PPC32 p440)
DIP("lfdx fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
putFReg( frD_addr, loadBE(Ity_F64, mkexpr(EA)) );
break;
case 0x277: // lfdux (Load Float Double, Update Indxd, PPC32 p439)
if (rA_addr == 0)
return False;
DIP("lfdux fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rA_idxd(rA_addr, rB_addr) );
putFReg( frD_addr, loadBE(Ity_F64, mkexpr(EA)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x357: // lfiwax (Load Float As Integer, Indxd, ISA 2.05 p120)
DIP("lfiwax fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
assign( iLo, loadBE(Ity_I32, mkexpr(EA)) );
assign( iHi, binop(Iop_Sub32,
mkU32(0),
binop(Iop_Shr32, mkexpr(iLo), mkU8(31))) );
putFReg( frD_addr, unop(Iop_ReinterpI64asF64,
binop(Iop_32HLto64, mkexpr(iHi), mkexpr(iLo))) );
break;
case 0x377: // lfiwzx (Load floating-point as integer word, zero indexed
{
IRTemp dw = newTemp( Ity_I64 );
DIP("lfiwzx fr%u,r%u,r%u\n", frD_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
assign( iLo, loadBE(Ity_I32, mkexpr(EA)) );
assign( dw, binop( Iop_32HLto64, mkU32( 0 ), mkexpr( iLo ) ) );
putFReg( frD_addr, unop( Iop_ReinterpI64asF64, mkexpr( dw ) ) );
break;
}
default:
vex_printf("dis_fp_load(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_fp_load(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Floating Point Store Instructions
*/
static Bool dis_fp_store ( UInt theInstr )
{
/* X-Form, D-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
Int uimm16 = ifieldUIMM16(theInstr);
Int simm16 = extend_s_16to32(uimm16);
IRTemp frS = newTemp(Ity_F64);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRTemp rA = newTemp(ty);
IRTemp rB = newTemp(ty);
assign( frS, getFReg(frS_addr) );
assign( rA, getIReg(rA_addr) );
assign( rB, getIReg(rB_addr) );
/* These are straightforward from a status bits perspective: no
funny status or CR bits affected. For single precision stores,
the values are truncated and denormalised (not rounded) to turn
them into single precision values. */
switch (opc1) {
case 0x34: // stfs (Store Float Single, PPC32 p518)
DIP("stfs fr%u,%d(r%u)\n", frS_addr, simm16, rA_addr);
assign( EA, ea_rAor0_simm(rA_addr, simm16) );
/* Use Iop_TruncF64asF32 to truncate and possible denormalise
the value to be stored in the correct way, without any
rounding. */
storeBE( mkexpr(EA),
unop(Iop_TruncF64asF32, mkexpr(frS)) );
break;
case 0x35: // stfsu (Store Float Single, Update, PPC32 p519)
if (rA_addr == 0)
return False;
DIP("stfsu fr%u,%d(r%u)\n", frS_addr, simm16, rA_addr);
assign( EA, ea_rA_simm(rA_addr, simm16) );
/* See comment for stfs */
storeBE( mkexpr(EA),
unop(Iop_TruncF64asF32, mkexpr(frS)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x36: // stfd (Store Float Double, PPC32 p513)
DIP("stfd fr%u,%d(r%u)\n", frS_addr, simm16, rA_addr);
assign( EA, ea_rAor0_simm(rA_addr, simm16) );
storeBE( mkexpr(EA), mkexpr(frS) );
break;
case 0x37: // stfdu (Store Float Double, Update, PPC32 p514)
if (rA_addr == 0)
return False;
DIP("stfdu fr%u,%d(r%u)\n", frS_addr, simm16, rA_addr);
assign( EA, ea_rA_simm(rA_addr, simm16) );
storeBE( mkexpr(EA), mkexpr(frS) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x1F:
if (b0 != 0) {
vex_printf("dis_fp_store(ppc)(instr,b0)\n");
return False;
}
switch(opc2) {
case 0x297: // stfsx (Store Float Single Indexed, PPC32 p521)
DIP("stfsx fr%u,r%u,r%u\n", frS_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
/* See note for stfs */
storeBE( mkexpr(EA),
unop(Iop_TruncF64asF32, mkexpr(frS)) );
break;
case 0x2B7: // stfsux (Store Float Sgl, Update Indxd, PPC32 p520)
if (rA_addr == 0)
return False;
DIP("stfsux fr%u,r%u,r%u\n", frS_addr, rA_addr, rB_addr);
assign( EA, ea_rA_idxd(rA_addr, rB_addr) );
/* See note for stfs */
storeBE( mkexpr(EA),
unop(Iop_TruncF64asF32, mkexpr(frS)) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x2D7: // stfdx (Store Float Double Indexed, PPC32 p516)
DIP("stfdx fr%u,r%u,r%u\n", frS_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
storeBE( mkexpr(EA), mkexpr(frS) );
break;
case 0x2F7: // stfdux (Store Float Dbl, Update Indxd, PPC32 p515)
if (rA_addr == 0)
return False;
DIP("stfdux fr%u,r%u,r%u\n", frS_addr, rA_addr, rB_addr);
assign( EA, ea_rA_idxd(rA_addr, rB_addr) );
storeBE( mkexpr(EA), mkexpr(frS) );
putIReg( rA_addr, mkexpr(EA) );
break;
case 0x3D7: // stfiwx (Store Float as Int, Indexed, PPC32 p517)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
DIP("stfiwx fr%u,r%u,r%u\n", frS_addr, rA_addr, rB_addr);
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
storeBE( mkexpr(EA),
unop(Iop_64to32, unop(Iop_ReinterpF64asI64, mkexpr(frS))) );
break;
default:
vex_printf("dis_fp_store(ppc)(opc2)\n");
return False;
}
break;
default:
vex_printf("dis_fp_store(ppc)(opc1)\n");
return False;
}
return True;
}
/*
Floating Point Arith Instructions
*/
static Bool dis_fp_arith ( UInt theInstr )
{
/* A-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frD_addr = ifieldRegDS(theInstr);
UChar frA_addr = ifieldRegA(theInstr);
UChar frB_addr = ifieldRegB(theInstr);
UChar frC_addr = ifieldRegC(theInstr);
UChar opc2 = ifieldOPClo5(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRTemp frD = newTemp(Ity_F64);
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
IRTemp frC = newTemp(Ity_F64);
IRExpr* rm = get_IR_roundingmode();
/* By default, we will examine the results of the operation and set
fpscr[FPRF] accordingly. */
Bool set_FPRF = True;
/* By default, if flag_RC is set, we will clear cr1 after the
operation. In reality we should set cr1 to indicate the
exception status of the operation, but since we're not
simulating exceptions, the exception status will appear to be
zero. Hence cr1 should be cleared if this is a . form insn. */
Bool clear_CR1 = True;
assign( frA, getFReg(frA_addr));
assign( frB, getFReg(frB_addr));
assign( frC, getFReg(frC_addr));
switch (opc1) {
case 0x3B:
switch (opc2) {
case 0x12: // fdivs (Floating Divide Single, PPC32 p407)
if (frC_addr != 0)
return False;
DIP("fdivs%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop( Iop_DivF64r32,
rm, mkexpr(frA), mkexpr(frB) ));
break;
case 0x14: // fsubs (Floating Subtract Single, PPC32 p430)
if (frC_addr != 0)
return False;
DIP("fsubs%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop( Iop_SubF64r32,
rm, mkexpr(frA), mkexpr(frB) ));
break;
case 0x15: // fadds (Floating Add Single, PPC32 p401)
if (frC_addr != 0)
return False;
DIP("fadds%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop( Iop_AddF64r32,
rm, mkexpr(frA), mkexpr(frB) ));
break;
case 0x16: // fsqrts (Floating SqRt (Single-Precision), PPC32 p428)
// NOTE: POWERPC OPTIONAL, "General-Purpose Group" (PPC32_FX)
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("fsqrts%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
// however illogically, on ppc970 this insn behaves identically
// to fsqrt (double-precision). So use SqrtF64, not SqrtF64r32.
assign( frD, binop( Iop_SqrtF64, rm, mkexpr(frB) ));
break;
case 0x18: // fres (Floating Reciprocal Estimate Single, PPC32 p421)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("fres%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
{ IRExpr* ieee_one
= IRExpr_Const(IRConst_F64i(0x3ff0000000000000ULL));
assign( frD, triop( Iop_DivF64r32,
rm,
ieee_one, mkexpr(frB) ));
}
break;
case 0x19: // fmuls (Floating Multiply Single, PPC32 p414)
if (frB_addr != 0)
return False;
DIP("fmuls%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr);
assign( frD, triop( Iop_MulF64r32,
rm, mkexpr(frA), mkexpr(frC) ));
break;
case 0x1A: // frsqrtes (Floating Recip SqRt Est Single)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
// Undocumented instruction?
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("frsqrtes%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
assign( frD, unop(Iop_Est5FRSqrt, mkexpr(frB)) );
break;
default:
vex_printf("dis_fp_arith(ppc)(3B: opc2)\n");
return False;
}
break;
case 0x3F:
switch (opc2) {
case 0x12: // fdiv (Floating Div (Double-Precision), PPC32 p406)
if (frC_addr != 0)
return False;
DIP("fdiv%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop(Iop_DivF64, rm, mkexpr(frA), mkexpr(frB)) );
break;
case 0x14: // fsub (Floating Sub (Double-Precision), PPC32 p429)
if (frC_addr != 0)
return False;
DIP("fsub%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop(Iop_SubF64, rm, mkexpr(frA), mkexpr(frB)) );
break;
case 0x15: // fadd (Floating Add (Double-Precision), PPC32 p400)
if (frC_addr != 0)
return False;
DIP("fadd%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frB_addr);
assign( frD, triop(Iop_AddF64, rm, mkexpr(frA), mkexpr(frB)) );
break;
case 0x16: // fsqrt (Floating SqRt (Double-Precision), PPC32 p427)
// NOTE: POWERPC OPTIONAL, "General-Purpose Group" (PPC32_FX)
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("fsqrt%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
assign( frD, binop(Iop_SqrtF64, rm, mkexpr(frB)) );
break;
case 0x17: { // fsel (Floating Select, PPC32 p426)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
IRTemp cc = newTemp(Ity_I32);
IRTemp cc_b0 = newTemp(Ity_I32);
DIP("fsel%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
// cc: UN == 0x41, LT == 0x01, GT == 0x00, EQ == 0x40
// => GT|EQ == (cc & 0x1 == 0)
assign( cc, binop(Iop_CmpF64, mkexpr(frA),
IRExpr_Const(IRConst_F64(0))) );
assign( cc_b0, binop(Iop_And32, mkexpr(cc), mkU32(1)) );
// frD = (frA >= 0.0) ? frC : frB
// = (cc_b0 == 0) ? frC : frB
assign( frD,
IRExpr_Mux0X(
unop(Iop_1Uto8,
binop(Iop_CmpEQ32, mkexpr(cc_b0), mkU32(0))),
mkexpr(frB),
mkexpr(frC) ));
/* One of the rare ones which don't mess with FPRF */
set_FPRF = False;
break;
}
case 0x18: // fre (Floating Reciprocal Estimate)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
// Note: unclear whether this insn really exists or not
// ppc970 doesn't have it, but POWER5 does
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("fre%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
{ IRExpr* ieee_one
= IRExpr_Const(IRConst_F64i(0x3ff0000000000000ULL));
assign( frD, triop( Iop_DivF64,
rm,
ieee_one, mkexpr(frB) ));
}
break;
case 0x19: // fmul (Floating Mult (Double Precision), PPC32 p413)
if (frB_addr != 0)
vex_printf("dis_fp_arith(ppc)(instr,fmul)\n");
DIP("fmul%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr);
assign( frD, triop(Iop_MulF64, rm, mkexpr(frA), mkexpr(frC)) );
break;
case 0x1A: // frsqrte (Floating Recip SqRt Est., PPC32 p424)
// NOTE: POWERPC OPTIONAL, "Graphics Group" (PPC32_GX)
if (frA_addr != 0 || frC_addr != 0)
return False;
DIP("frsqrte%s fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frB_addr);
assign( frD, unop(Iop_Est5FRSqrt, mkexpr(frB)) );
break;
default:
vex_printf("dis_fp_arith(ppc)(3F: opc2)\n");
return False;
}
break;
default:
vex_printf("dis_fp_arith(ppc)(opc1)\n");
return False;
}
putFReg( frD_addr, mkexpr(frD) );
if (set_FPRF) {
// XXX XXX XXX FIXME
// set FPRF from frD
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8(0) );
putCR0( 1, mkU8(0) );
}
return True;
}
/*
Floating Point Mult-Add Instructions
*/
static Bool dis_fp_multadd ( UInt theInstr )
{
/* A-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frD_addr = ifieldRegDS(theInstr);
UChar frA_addr = ifieldRegA(theInstr);
UChar frB_addr = ifieldRegB(theInstr);
UChar frC_addr = ifieldRegC(theInstr);
UChar opc2 = ifieldOPClo5(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRTemp frD = newTemp(Ity_F64);
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
IRTemp frC = newTemp(Ity_F64);
IRTemp rmt = newTemp(Ity_I32);
IRExpr* rm;
/* By default, we will examine the results of the operation and set
fpscr[FPRF] accordingly. */
Bool set_FPRF = True;
/* By default, if flag_RC is set, we will clear cr1 after the
operation. In reality we should set cr1 to indicate the
exception status of the operation, but since we're not
simulating exceptions, the exception status will appear to be
zero. Hence cr1 should be cleared if this is a . form insn. */
Bool clear_CR1 = True;
/* Bind the rounding mode expression to a temp; there's no
point in creating gratuitous CSEs, as we know we'll need
to use it twice. */
assign( rmt, get_IR_roundingmode() );
rm = mkexpr(rmt);
assign( frA, getFReg(frA_addr));
assign( frB, getFReg(frB_addr));
assign( frC, getFReg(frC_addr));
/* The rounding in this is all a bit dodgy. The idea is to only do
one rounding. That clearly isn't achieveable without dedicated
four-input IR primops, although in the single precision case we
can sort-of simulate it by doing the inner multiply in double
precision.
In the negated cases, the negation happens after rounding. */
switch (opc1) {
case 0x3B:
switch (opc2) {
case 0x1C: // fmsubs (Floating Mult-Subtr Single, PPC32 p412)
DIP("fmsubs%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, qop( Iop_MSubF64r32, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) ));
break;
case 0x1D: // fmadds (Floating Mult-Add Single, PPC32 p409)
DIP("fmadds%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, qop( Iop_MAddF64r32, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) ));
break;
case 0x1E: // fnmsubs (Float Neg Mult-Subtr Single, PPC32 p420)
DIP("fnmsubs%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, unop( Iop_NegF64,
qop( Iop_MSubF64r32, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) )));
break;
case 0x1F: // fnmadds (Floating Negative Multiply-Add Single, PPC32 p418)
DIP("fnmadds%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, unop( Iop_NegF64,
qop( Iop_MAddF64r32, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) )));
break;
default:
vex_printf("dis_fp_multadd(ppc)(3B: opc2)\n");
return False;
}
break;
case 0x3F:
switch (opc2) {
case 0x1C: // fmsub (Float Mult-Sub (Dbl Precision), PPC32 p411)
DIP("fmsub%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, qop( Iop_MSubF64, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) ));
break;
case 0x1D: // fmadd (Float Mult-Add (Dbl Precision), PPC32 p408)
DIP("fmadd%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, qop( Iop_MAddF64, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) ));
break;
case 0x1E: // fnmsub (Float Neg Mult-Subtr (Dbl Precision), PPC32 p419)
DIP("fnmsub%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, unop( Iop_NegF64,
qop( Iop_MSubF64, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) )));
break;
case 0x1F: // fnmadd (Float Neg Mult-Add (Dbl Precision), PPC32 p417)
DIP("fnmadd%s fr%u,fr%u,fr%u,fr%u\n", flag_rC ? ".":"",
frD_addr, frA_addr, frC_addr, frB_addr);
assign( frD, unop( Iop_NegF64,
qop( Iop_MAddF64, rm,
mkexpr(frA), mkexpr(frC), mkexpr(frB) )));
break;
default:
vex_printf("dis_fp_multadd(ppc)(3F: opc2)\n");
return False;
}
break;
default:
vex_printf("dis_fp_multadd(ppc)(opc1)\n");
return False;
}
putFReg( frD_addr, mkexpr(frD) );
if (set_FPRF) {
// XXX XXX XXX FIXME
// set FPRF from frD
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8(0) );
putCR0( 1, mkU8(0) );
}
return True;
}
/*
* fe_flag is set to 1 if any of the following conditions occurs:
* - The floating-point operand in register FRB is a Zero, a
* NaN, an Infinity, or a negative value.
* - e_b is less than or equal to: -970 for double precision; -103 for single precision
* Otherwise fe_flag is set to 0.
*
* fg_flag is set to 1 if either of the following conditions occurs.
* - The floating-point operand in register FRB is a Zero, an
* Infinity, or a denormalized value.
* Otherwise fg_flag is set to 0.
*
*/
static void do_fp_tsqrt(IRTemp frB_Int, Bool sp, IRTemp * fe_flag_tmp, IRTemp * fg_flag_tmp)
{
// The following temps are for holding intermediate results
IRTemp e_b = newTemp(Ity_I32);
IRExpr * fe_flag, * fg_flag;
IRTemp frB_exp_shR = newTemp(Ity_I32);
UInt bias = sp? 127 : 1023;
IRExpr * frbNaN, * frbDenorm, * frBNeg;
IRExpr * eb_LTE;
IRTemp frbZero_tmp = newTemp(Ity_I1);
IRTemp frbInf_tmp = newTemp(Ity_I1);
*fe_flag_tmp = newTemp(Ity_I32);
*fg_flag_tmp = newTemp(Ity_I32);
assign( frB_exp_shR, fp_exp_part( frB_Int, sp ) );
assign(e_b, binop( Iop_Sub32, mkexpr(frB_exp_shR), mkU32( bias ) ));
////////////////// fe_flag tests BEGIN //////////////////////
/* We first do all tests that may result in setting fe_flag to '1'.
* (NOTE: These tests are similar to those used for ftdiv. See do_fp_tdiv()
* for details.)
*/
frbNaN = sp ? is_NaN_32(frB_Int) : is_NaN(frB_Int);
assign( frbInf_tmp, is_Inf(frB_Int, sp) );
assign( frbZero_tmp, is_Zero(frB_Int, sp ) );
{
// Test_value = -970 for double precision
UInt test_value = sp ? 0xffffff99 : 0xfffffc36;
eb_LTE = binop( Iop_CmpLE32S, mkexpr( e_b ), mkU32( test_value ) );
}
frBNeg = binop( Iop_CmpEQ32,
binop( Iop_Shr32,
sp ? mkexpr( frB_Int ) : unop( Iop_64HIto32, mkexpr( frB_Int ) ),
mkU8( 31 ) ),
mkU32( 1 ) );
////////////////// fe_flag tests END //////////////////////
////////////////// fg_flag tests BEGIN //////////////////////
/*
* The following tests were already performed above in the fe_flag
* tests. So these conditions will result in both fe_ and fg_ flags
* being set.
* - Test if FRB is Zero
* - Test if FRB is an Infinity
*/
/*
* Test if FRB holds a denormalized value. A denormalized value is one where
* the exp is 0 and the fraction is non-zero.
*/
if (sp) {
IRTemp frac_part = newTemp(Ity_I32);
assign( frac_part, binop( Iop_And32, mkexpr(frB_Int), mkU32(0x007fffff)) );
frbDenorm
= mkAND1( binop( Iop_CmpEQ32, mkexpr( frB_exp_shR ), mkU32( 0 ) ),
binop( Iop_CmpNE32, mkexpr( frac_part ), mkU32( 0 ) ) );
} else {
IRExpr * hi32, * low32, * fraction_is_nonzero;
IRTemp frac_part = newTemp(Ity_I64);
assign( frac_part, FP_FRAC_PART(frB_Int) );
hi32 = unop( Iop_64HIto32, mkexpr( frac_part ) );
low32 = unop( Iop_64to32, mkexpr( frac_part ) );
fraction_is_nonzero = binop( Iop_CmpNE32, binop( Iop_Or32, low32, hi32 ),
mkU32( 0 ) );
frbDenorm
= mkAND1( binop( Iop_CmpEQ32, mkexpr( frB_exp_shR ), mkU32( 0 ) ),
fraction_is_nonzero );
}
////////////////// fg_flag tests END //////////////////////
/////////////////////////
fe_flag = mkOR1( mkexpr( frbZero_tmp ),
mkOR1( frbNaN,
mkOR1( mkexpr( frbInf_tmp ),
mkOR1( frBNeg, eb_LTE ) ) ) );
fe_flag = unop(Iop_1Uto32, fe_flag);
fg_flag = mkOR1( mkexpr( frbZero_tmp ),
mkOR1( mkexpr( frbInf_tmp ), frbDenorm ) );
fg_flag = unop(Iop_1Uto32, fg_flag);
assign (*fg_flag_tmp, fg_flag);
assign (*fe_flag_tmp, fe_flag);
}
/*
* fe_flag is set to 1 if any of the following conditions occurs:
* - The double-precision floating-point operand in register FRA is a NaN or an
* Infinity.
* - The double-precision floating-point operand in register FRB is a Zero, a
* NaN, or an Infinity.
* - e_b is less than or equal to -1022.
* - e_b is greater than or equal to 1021.
* - The double-precision floating-point operand in register FRA is not a zero
* and the difference, e_a - e_b, is greater than or equal to 1023.
* - The double-precision floating-point operand in register FRA is not a zero
* and the difference, e_a - e_b, is less than or equal to -1021.
* - The double-precision floating-point operand in register FRA is not a zero
* and e_a is less than or equal to -970
* Otherwise fe_flag is set to 0.
*
* fg_flag is set to 1 if either of the following conditions occurs.
* - The double-precision floating-point operand in register FRA is an Infinity.
* - The double-precision floating-point operand in register FRB is a Zero, an
* Infinity, or a denormalized value.
* Otherwise fg_flag is set to 0.
*
*/
static void _do_fp_tdiv(IRTemp frA_int, IRTemp frB_int, Bool sp, IRTemp * fe_flag_tmp, IRTemp * fg_flag_tmp)
{
// The following temps are for holding intermediate results
IRTemp e_a = newTemp(Ity_I32);
IRTemp e_b = newTemp(Ity_I32);
IRTemp frA_exp_shR = newTemp(Ity_I32);
IRTemp frB_exp_shR = newTemp(Ity_I32);
UInt bias = sp? 127 : 1023;
*fe_flag_tmp = newTemp(Ity_I32);
*fg_flag_tmp = newTemp(Ity_I32);
/* The following variables hold boolean results from tests
* that are OR'ed together for setting the fe_ and fg_ flags.
* For some cases, the booleans are used more than once, so
* I make those IRTemp's instead of IRExpr's.
*/
IRExpr * fraNaN, * frbNaN, * frbDenorm;
IRExpr * eb_LTE, * eb_GTE, * ea_eb_GTE, * ea_eb_LTE, * ea_LTE;
IRTemp fraInf_tmp = newTemp(Ity_I1);
IRTemp frbZero_tmp = newTemp(Ity_I1);
IRTemp frbInf_tmp = newTemp(Ity_I1);
IRTemp fraNotZero_tmp = newTemp(Ity_I1);
/* The following are the flags that are set by OR'ing the results of
* all the tests done for tdiv. These flags are the input to the specified CR.
*/
IRExpr * fe_flag, * fg_flag;
// Create temps that will be used throughout the following tests.
assign( frA_exp_shR, fp_exp_part( frA_int, sp ) );
assign( frB_exp_shR, fp_exp_part( frB_int, sp ) );
/* Let e_[a|b] be the unbiased exponent: i.e. exp - 1023. */
assign(e_a, binop( Iop_Sub32, mkexpr(frA_exp_shR), mkU32( bias ) ));
assign(e_b, binop( Iop_Sub32, mkexpr(frB_exp_shR), mkU32( bias ) ));
////////////////// fe_flag tests BEGIN //////////////////////
/* We first do all tests that may result in setting fe_flag to '1'. */
/*
* Test if the double-precision floating-point operand in register FRA is
* a NaN:
*/
fraNaN = sp ? is_NaN_32(frA_int) : is_NaN(frA_int);
/*
* Test if the double-precision floating-point operand in register FRA is
* an Infinity.
*/
assign(fraInf_tmp, is_Inf(frA_int, sp));
/*
* Test if the double-precision floating-point operand in register FRB is
* a NaN:
*/
frbNaN = sp ? is_NaN_32(frB_int) : is_NaN(frB_int);
/*
* Test if the double-precision floating-point operand in register FRB is
* an Infinity.
*/
assign( frbInf_tmp, is_Inf(frB_int, sp) );
/*
* Test if the double-precision floating-point operand in register FRB is
* a Zero.
*/
assign( frbZero_tmp, is_Zero(frB_int, sp) );
/*
* Test if e_b <= -1022 for double precision;
* or e_b <= -126 for single precision
*/
{
UInt test_value = sp ? 0xffffff82 : 0xfffffc02;
eb_LTE = binop(Iop_CmpLE32S, mkexpr(e_b), mkU32(test_value));
}
/*
* Test if e_b >= 1021 (i.e., 1021 < e_b) for double precision;
* or e_b >= -125 (125 < e_b) for single precision
*/
{
Int test_value = sp ? 125 : 1021;
eb_GTE = binop(Iop_CmpLT32S, mkU32(test_value), mkexpr(e_b));
}
/*
* Test if FRA != Zero and (e_a - e_b) >= bias
*/
assign( fraNotZero_tmp, unop( Iop_Not1, is_Zero( frA_int, sp ) ) );
ea_eb_GTE = mkAND1( mkexpr( fraNotZero_tmp ),
binop( Iop_CmpLT32S, mkU32( bias ),
binop( Iop_Sub32, mkexpr( e_a ),
mkexpr( e_b ) ) ) );
/*
* Test if FRA != Zero and (e_a - e_b) <= [-1021 (double precision) or -125 (single precision)]
*/
{
UInt test_value = sp ? 0xffffff83 : 0xfffffc03;
ea_eb_LTE = mkAND1( mkexpr( fraNotZero_tmp ),
binop( Iop_CmpLE32S,
binop( Iop_Sub32,
mkexpr( e_a ),
mkexpr( e_b ) ),
mkU32( test_value ) ) );
}
/*
* Test if FRA != Zero and e_a <= [-970 (double precision) or -103 (single precision)]
*/
{
UInt test_value = 0xfffffc36; //Int test_value = -970;
ea_LTE = mkAND1( mkexpr( fraNotZero_tmp ), binop( Iop_CmpLE32S,
mkexpr( e_a ),
mkU32( test_value ) ) );
}
////////////////// fe_flag tests END //////////////////////
////////////////// fg_flag tests BEGIN //////////////////////
/*
* The following tests were already performed above in the fe_flag
* tests. So these conditions will result in both fe_ and fg_ flags
* being set.
* - Test if FRA is an Infinity
* - Test if FRB ix Zero
* - Test if FRB is an Infinity
*/
/*
* Test if FRB holds a denormalized value. A denormalized value is one where
* the exp is 0 and the fraction is non-zero.
*/
{
IRExpr * fraction_is_nonzero;
if (sp) {
fraction_is_nonzero = binop( Iop_CmpNE32, FP_FRAC_PART32(frB_int),
mkU32( 0 ) );
} else {
IRExpr * hi32, * low32;
IRTemp frac_part = newTemp(Ity_I64);
assign( frac_part, FP_FRAC_PART(frB_int) );
hi32 = unop( Iop_64HIto32, mkexpr( frac_part ) );
low32 = unop( Iop_64to32, mkexpr( frac_part ) );
fraction_is_nonzero = binop( Iop_CmpNE32, binop( Iop_Or32, low32, hi32 ),
mkU32( 0 ) );
}
frbDenorm = mkAND1( binop( Iop_CmpEQ32, mkexpr( frB_exp_shR ),
mkU32( 0x0 ) ), fraction_is_nonzero );
}
////////////////// fg_flag tests END //////////////////////
fe_flag
= mkOR1(
fraNaN,
mkOR1(
mkexpr( fraInf_tmp ),
mkOR1(
mkexpr( frbZero_tmp ),
mkOR1(
frbNaN,
mkOR1(
mkexpr( frbInf_tmp ),
mkOR1( eb_LTE,
mkOR1( eb_GTE,
mkOR1( ea_eb_GTE,
mkOR1( ea_eb_LTE,
ea_LTE ) ) ) ) ) ) ) ) );
fe_flag = unop(Iop_1Uto32, fe_flag);
fg_flag = mkOR1( mkexpr( fraInf_tmp ), mkOR1( mkexpr( frbZero_tmp ),
mkOR1( mkexpr( frbInf_tmp ),
frbDenorm ) ) );
fg_flag = unop(Iop_1Uto32, fg_flag);
assign(*fe_flag_tmp, fe_flag);
assign(*fg_flag_tmp, fg_flag);
}
/* See description for _do_fp_tdiv() above. */
static IRExpr * do_fp_tdiv(IRTemp frA_int, IRTemp frB_int)
{
IRTemp fe_flag, fg_flag;
/////////////////////////
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
IRExpr * fl_flag = unop(Iop_Not32, mkU32(0xFFFFFE));
fe_flag = fg_flag = IRTemp_INVALID;
_do_fp_tdiv(frA_int, frB_int, False/*not single precision*/, &fe_flag, &fg_flag);
return binop( Iop_Or32,
binop( Iop_Or32,
binop( Iop_Shl32, fl_flag, mkU8( 3 ) ),
binop( Iop_Shl32, mkexpr(fg_flag), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag), mkU8( 1 ) ) );
}
static Bool dis_fp_tests ( UInt theInstr )
{
UChar opc1 = ifieldOPC(theInstr);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
UChar frB_addr = ifieldRegB(theInstr);
UChar b0 = ifieldBIT0(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
IRTemp frB_I64 = newTemp(Ity_I64);
if (opc1 != 0x3F || b0 != 0 ){
vex_printf("dis_fp_tests(ppc)(ftdiv)\n");
return False;
}
assign( frB_I64, unop( Iop_ReinterpF64asI64, getFReg( frB_addr ) ) );
switch (opc2) {
case 0x080: // ftdiv
{
UChar frA_addr = ifieldRegA(theInstr);
IRTemp frA_I64 = newTemp(Ity_I64);
UChar b21to22 = toUChar( IFIELD( theInstr, 21, 2 ) );
if (b21to22 != 0 ) {
vex_printf("dis_fp_tests(ppc)(ftdiv)\n");
return False;
}
assign( frA_I64, unop( Iop_ReinterpF64asI64, getFReg( frA_addr ) ) );
putGST_field( PPC_GST_CR, do_fp_tdiv(frA_I64, frB_I64), crfD );
DIP("ftdiv crf%d,fr%u,fr%u\n", crfD, frA_addr, frB_addr);
break;
}
case 0x0A0: // ftsqrt
{
IRTemp flags = newTemp(Ity_I32);
IRTemp fe_flag, fg_flag;
fe_flag = fg_flag = IRTemp_INVALID;
UChar b18to22 = toUChar( IFIELD( theInstr, 18, 5 ) );
if ( b18to22 != 0) {
vex_printf("dis_fp_tests(ppc)(ftsqrt)\n");
return False;
}
DIP("ftsqrt crf%d,fr%u\n", crfD, frB_addr);
do_fp_tsqrt(frB_I64, False /* not single precision*/, &fe_flag, &fg_flag);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flags,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR, mkexpr(flags), crfD );
break;
}
default:
vex_printf("dis_fp_tests(ppc)(opc2)\n");
return False;
}
return True;
}
/*
Floating Point Compare Instructions
*/
static Bool dis_fp_cmp ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
UChar b21to22 = toUChar( IFIELD( theInstr, 21, 2 ) );
UChar frA_addr = ifieldRegA(theInstr);
UChar frB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRTemp ccIR = newTemp(Ity_I32);
IRTemp ccPPC32 = newTemp(Ity_I32);
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
if (opc1 != 0x3F || b21to22 != 0 || b0 != 0) {
vex_printf("dis_fp_cmp(ppc)(instr)\n");
return False;
}
assign( frA, getFReg(frA_addr));
assign( frB, getFReg(frB_addr));
assign( ccIR, binop(Iop_CmpF64, mkexpr(frA), mkexpr(frB)) );
/* Map compare result from IR to PPC32 */
/*
FP cmp result | PPC | IR
--------------------------
UN | 0x1 | 0x45
EQ | 0x2 | 0x40
GT | 0x4 | 0x00
LT | 0x8 | 0x01
*/
// ccPPC32 = Shl(1, (~(ccIR>>5) & 2)
// | ((ccIR ^ (ccIR>>6)) & 1)
assign(
ccPPC32,
binop(
Iop_Shl32,
mkU32(1),
unop(
Iop_32to8,
binop(
Iop_Or32,
binop(
Iop_And32,
unop(
Iop_Not32,
binop(Iop_Shr32, mkexpr(ccIR), mkU8(5))
),
mkU32(2)
),
binop(
Iop_And32,
binop(
Iop_Xor32,
mkexpr(ccIR),
binop(Iop_Shr32, mkexpr(ccIR), mkU8(6))
),
mkU32(1)
)
)
)
)
);
putGST_field( PPC_GST_CR, mkexpr(ccPPC32), crfD );
/* CAB: TODO?: Support writing cc to FPSCR->FPCC ?
putGST_field( PPC_GST_FPSCR, mkexpr(ccPPC32), 4 );
*/
// XXX XXX XXX FIXME
// Also write the result into FPRF (it's not entirely clear how)
/* Note: Differences between fcmpu and fcmpo are only in exception
flag settings, which aren't supported anyway. */
switch (opc2) {
case 0x000: // fcmpu (Floating Compare Unordered, PPC32 p403)
DIP("fcmpu crf%d,fr%u,fr%u\n", crfD, frA_addr, frB_addr);
break;
case 0x020: // fcmpo (Floating Compare Ordered, PPC32 p402)
DIP("fcmpo crf%d,fr%u,fr%u\n", crfD, frA_addr, frB_addr);
break;
default:
vex_printf("dis_fp_cmp(ppc)(opc2)\n");
return False;
}
return True;
}
/*
Floating Point Rounding/Conversion Instructions
*/
static Bool dis_fp_round ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar b16to20 = ifieldRegA(theInstr);
UChar frD_addr = ifieldRegDS(theInstr);
UChar frB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRTemp frD = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
IRTemp r_tmp32 = newTemp(Ity_I32);
IRTemp r_tmp64 = newTemp(Ity_I64);
IRExpr* rm = get_IR_roundingmode();
/* By default, we will examine the results of the operation and set
fpscr[FPRF] accordingly. */
Bool set_FPRF = True;
/* By default, if flag_RC is set, we will clear cr1 after the
operation. In reality we should set cr1 to indicate the
exception status of the operation, but since we're not
simulating exceptions, the exception status will appear to be
zero. Hence cr1 should be cleared if this is a . form insn. */
Bool clear_CR1 = True;
if ((!(opc1 == 0x3F || opc1 == 0x3B)) || b16to20 != 0) {
vex_printf("dis_fp_round(ppc)(instr)\n");
return False;
}
assign( frB, getFReg(frB_addr));
if (opc1 == 0x3B) {
/* The fcfid[u]s instructions (from ISA 2.06) are a bit odd because
* they're very similar to the other instructions handled here, but have
* a different primary opcode.
*/
switch (opc2) {
case 0x34E: // fcfids (Float convert from signed DWord to single precision)
DIP("fcfids%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64, unop( Iop_ReinterpF64asI64, mkexpr(frB)) );
assign( frD, binop( Iop_RoundF64toF32, rm, binop( Iop_I64StoF64, rm,
mkexpr( r_tmp64 ) ) ) );
goto putFR;
case 0x3Ce: // fcfidus (Float convert from unsigned DWord to single precision)
DIP("fcfidus%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64, unop( Iop_ReinterpF64asI64, mkexpr(frB)) );
assign( frD, unop( Iop_F32toF64, binop( Iop_I64UtoF32, rm, mkexpr( r_tmp64 ) ) ) );
goto putFR;
}
}
switch (opc2) {
case 0x00C: // frsp (Float Round to Single, PPC32 p423)
DIP("frsp%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( frD, binop( Iop_RoundF64toF32, rm, mkexpr(frB) ));
break;
case 0x00E: // fctiw (Float Conv to Int, PPC32 p404)
DIP("fctiw%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp32,
binop(Iop_F64toI32S, rm, mkexpr(frB)) );
assign( frD, unop( Iop_ReinterpI64asF64,
unop( Iop_32Uto64, mkexpr(r_tmp32))));
/* FPRF is undefined after fctiw. Leave unchanged. */
set_FPRF = False;
break;
case 0x00F: // fctiwz (Float Conv to Int, Round to Zero, PPC32 p405)
DIP("fctiwz%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp32,
binop(Iop_F64toI32S, mkU32(Irrm_ZERO), mkexpr(frB) ));
assign( frD, unop( Iop_ReinterpI64asF64,
unop( Iop_32Uto64, mkexpr(r_tmp32))));
/* FPRF is undefined after fctiwz. Leave unchanged. */
set_FPRF = False;
break;
case 0x08F: case 0x08E: // fctiwu[z]
DIP("fctiwu%s%s fr%u,fr%u\n", opc2 == 0x08F ? "z" : "",
flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp32,
binop( Iop_F64toI32U,
opc2 == 0x08F ? mkU32( Irrm_ZERO ) : rm,
mkexpr( frB ) ) );
assign( frD, unop( Iop_ReinterpI64asF64,
unop( Iop_32Uto64, mkexpr(r_tmp32))));
/* FPRF is undefined after fctiwz. Leave unchanged. */
set_FPRF = False;
break;
case 0x32E: // fctid (Float Conv to Int DWord, PPC64 p437)
DIP("fctid%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, rm, mkexpr(frB)) );
assign( frD, unop( Iop_ReinterpI64asF64, mkexpr(r_tmp64)) );
/* FPRF is undefined after fctid. Leave unchanged. */
set_FPRF = False;
break;
case 0x32F: // fctidz (Float Conv to Int DWord, Round to Zero, PPC64 p437)
DIP("fctidz%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, mkU32(Irrm_ZERO), mkexpr(frB)) );
assign( frD, unop( Iop_ReinterpI64asF64, mkexpr(r_tmp64)) );
/* FPRF is undefined after fctidz. Leave unchanged. */
set_FPRF = False;
break;
case 0x3AE: case 0x3AF: // fctidu[z] (Float Conv to Int DWord Unsigned [Round to Zero])
{
DIP("fctidu%s%s fr%u,fr%u\n", opc2 == 0x3AE ? "" : "z",
flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64U, opc2 == 0x3AE ? rm : mkU32(Irrm_ZERO), mkexpr(frB)) );
assign( frD, unop( Iop_ReinterpI64asF64, mkexpr(r_tmp64)) );
/* FPRF is undefined after fctidz. Leave unchanged. */
set_FPRF = False;
break;
}
case 0x34E: // fcfid (Float Conv from Int DWord, PPC64 p434)
DIP("fcfid%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64, unop( Iop_ReinterpF64asI64, mkexpr(frB)) );
assign( frD,
binop(Iop_I64StoF64, rm, mkexpr(r_tmp64)) );
break;
case 0x3CE: // fcfidu (Float convert from unsigned DWord)
DIP("fcfidu%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64, unop( Iop_ReinterpF64asI64, mkexpr(frB)) );
assign( frD, binop( Iop_I64UtoF64, rm, mkexpr( r_tmp64 ) ) );
break;
case 0x188: case 0x1A8: case 0x1C8: case 0x1E8: // frin, friz, frip, frim
switch(opc2) {
case 0x188: // frin (Floating Round to Integer Nearest)
DIP("frin%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, mkU32(Irrm_NEAREST), mkexpr(frB)) );
break;
case 0x1A8: // friz (Floating Round to Integer Toward Zero)
DIP("friz%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, mkU32(Irrm_ZERO), mkexpr(frB)) );
break;
case 0x1C8: // frip (Floating Round to Integer Plus)
DIP("frip%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, mkU32(Irrm_PosINF), mkexpr(frB)) );
break;
case 0x1E8: // frim (Floating Round to Integer Minus)
DIP("frim%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( r_tmp64,
binop(Iop_F64toI64S, mkU32(Irrm_NegINF), mkexpr(frB)) );
break;
}
/* don't use the rounded integer if frB is outside -9e18..9e18 */
/* F64 has only log10(2**52) significant digits anyway */
/* need to preserve sign of zero */
/* frD = (fabs(frB) > 9e18) ? frB :
(sign(frB)) ? -fabs((double)r_tmp64) : (double)r_tmp64 */
assign(frD, IRExpr_Mux0X( unop(Iop_32to8,
binop(Iop_CmpF64,
IRExpr_Const(IRConst_F64(9e18)),
unop(Iop_AbsF64, mkexpr(frB)))),
IRExpr_Mux0X(unop(Iop_32to8,
binop(Iop_Shr32,
unop(Iop_64HIto32,
unop(Iop_ReinterpF64asI64,
mkexpr(frB))), mkU8(31))),
binop(Iop_I64StoF64, mkU32(0), mkexpr(r_tmp64) ),
unop(Iop_NegF64,
unop( Iop_AbsF64,
binop(Iop_I64StoF64, mkU32(0),
mkexpr(r_tmp64)) )) ),
mkexpr(frB)));
break;
default:
vex_printf("dis_fp_round(ppc)(opc2)\n");
return False;
}
putFR:
putFReg( frD_addr, mkexpr(frD) );
if (set_FPRF) {
// XXX XXX XXX FIXME
// set FPRF from frD
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8(0) );
putCR0( 1, mkU8(0) );
}
return True;
}
/*
Floating Point Pair Instructions
*/
static Bool dis_fp_pair ( UInt theInstr )
{
/* X-Form/DS-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frT_hi_addr = ifieldRegDS(theInstr);
UChar frT_lo_addr = frT_hi_addr + 1;
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt uimm16 = ifieldUIMM16(theInstr);
Int simm16 = extend_s_16to32(uimm16);
UInt opc2 = ifieldOPClo10(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA_hi = newTemp(ty);
IRTemp EA_lo = newTemp(ty);
IRTemp frT_hi = newTemp(Ity_F64);
IRTemp frT_lo = newTemp(Ity_F64);
UChar b0 = ifieldBIT0(theInstr);
Bool is_load = 0;
if ((frT_hi_addr %2) != 0) {
vex_printf("dis_fp_pair(ppc) : odd frT register\n");
return False;
}
switch (opc1) {
case 0x1F: // register offset
switch(opc2) {
case 0x317: // lfdpx (FP Load Double Pair X-form, ISA 2.05 p125)
DIP("ldpx fr%u,r%u,r%u\n", frT_hi_addr, rA_addr, rB_addr);
is_load = 1;
break;
case 0x397: // stfdpx (FP STORE Double Pair X-form, ISA 2.05 p125)
DIP("stdpx fr%u,r%u,r%u\n", frT_hi_addr, rA_addr, rB_addr);
break;
default:
vex_printf("dis_fp_pair(ppc) : X-form wrong opc2\n");
return False;
}
if (b0 != 0) {
vex_printf("dis_fp_pair(ppc)(0x1F,b0)\n");
return False;
}
assign( EA_hi, ea_rAor0_idxd( rA_addr, rB_addr ) );
break;
case 0x39: // lfdp (FP Load Double Pair DS-form, ISA 2.05 p125)
DIP("lfdp fr%u,%d(r%u)\n", frT_hi_addr, simm16, rA_addr);
assign( EA_hi, ea_rAor0_simm( rA_addr, simm16 ) );
is_load = 1;
break;
case 0x3d: // stfdp (FP Store Double Pair DS-form, ISA 2.05 p125)
DIP("stfdp fr%u,%d(r%u)\n", frT_hi_addr, simm16, rA_addr);
assign( EA_hi, ea_rAor0_simm( rA_addr, simm16 ) );
break;
default: // immediate offset
vex_printf("dis_fp_pair(ppc)(instr)\n");
return False;
}
if (mode64)
assign( EA_lo, binop(Iop_Add64, mkexpr(EA_hi), mkU64(8)) );
else
assign( EA_lo, binop(Iop_Add32, mkexpr(EA_hi), mkU32(8)) );
assign( frT_hi, getFReg(frT_hi_addr) );
assign( frT_lo, getFReg(frT_lo_addr) );
if (is_load) {
putFReg( frT_hi_addr, loadBE(Ity_F64, mkexpr(EA_hi)) );
putFReg( frT_lo_addr, loadBE(Ity_F64, mkexpr(EA_lo)) );
} else {
storeBE( mkexpr(EA_hi), mkexpr(frT_hi) );
storeBE( mkexpr(EA_lo), mkexpr(frT_lo) );
}
return True;
}
/*
Floating Point Move Instructions
*/
static Bool dis_fp_move ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar frD_addr = ifieldRegDS(theInstr);
UChar frA_addr = ifieldRegA(theInstr);
UChar frB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
IRTemp frD = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
IRTemp itmpB = newTemp(Ity_F64);
IRTemp frA;
IRTemp signA;
IRTemp hiD;
if (opc1 != 0x3F || (frA_addr != 0 && opc2 != 0x008)) {
vex_printf("dis_fp_move(ppc)(instr)\n");
return False;
}
assign( frB, getFReg(frB_addr));
switch (opc2) {
case 0x008: // fcpsgn (Floating Copy Sign, ISA_V2.05 p126)
DIP("fcpsgn%s fr%u,fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frA_addr,
frB_addr);
signA = newTemp(Ity_I32);
hiD = newTemp(Ity_I32);
itmpB = newTemp(Ity_I64);
frA = newTemp(Ity_F64);
assign( frA, getFReg(frA_addr) );
/* get A's sign bit */
assign(signA, binop(Iop_And32,
unop(Iop_64HIto32, unop(Iop_ReinterpF64asI64,
mkexpr(frA))),
mkU32(0x80000000)) );
assign( itmpB, unop(Iop_ReinterpF64asI64, mkexpr(frB)) );
/* mask off B's sign bit and or in A's sign bit */
assign(hiD, binop(Iop_Or32,
binop(Iop_And32,
unop(Iop_64HIto32,
mkexpr(itmpB)), /* frB's high 32 bits */
mkU32(0x7fffffff)),
mkexpr(signA)) );
/* combine hiD/loB into frD */
assign( frD, unop(Iop_ReinterpI64asF64,
binop(Iop_32HLto64,
mkexpr(hiD),
unop(Iop_64to32,
mkexpr(itmpB)))) ); /* frB's low 32 bits */
break;
case 0x028: // fneg (Floating Negate, PPC32 p416)
DIP("fneg%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( frD, unop( Iop_NegF64, mkexpr(frB) ));
break;
case 0x048: // fmr (Floating Move Register, PPC32 p410)
DIP("fmr%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( frD, mkexpr(frB) );
break;
case 0x088: // fnabs (Floating Negative Absolute Value, PPC32 p415)
DIP("fnabs%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( frD, unop( Iop_NegF64, unop( Iop_AbsF64, mkexpr(frB) )));
break;
case 0x108: // fabs (Floating Absolute Value, PPC32 p399)
DIP("fabs%s fr%u,fr%u\n", flag_rC ? ".":"", frD_addr, frB_addr);
assign( frD, unop( Iop_AbsF64, mkexpr(frB) ));
break;
default:
vex_printf("dis_fp_move(ppc)(opc2)\n");
return False;
}
putFReg( frD_addr, mkexpr(frD) );
/* None of these change FPRF. cr1 is set in the usual way though,
if flag_rC is set. */
if (flag_rC) {
putCR321( 1, mkU8(0) );
putCR0( 1, mkU8(0) );
}
return True;
}
/*
Floating Point Status/Control Register Instructions
*/
static Bool dis_fp_scr ( UInt theInstr, Bool GX_level )
{
/* Many forms - see each switch case */
UChar opc1 = ifieldOPC(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar flag_rC = ifieldBIT0(theInstr);
if (opc1 != 0x3F) {
vex_printf("dis_fp_scr(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x026: { // mtfsb1 (Move to FPSCR Bit 1, PPC32 p479)
// Bit crbD of the FPSCR is set.
UChar crbD = ifieldRegDS(theInstr);
UInt b11to20 = IFIELD(theInstr, 11, 10);
if (b11to20 != 0) {
vex_printf("dis_fp_scr(ppc)(instr,mtfsb1)\n");
return False;
}
DIP("mtfsb1%s crb%d \n", flag_rC ? ".":"", crbD);
putGST_masked( PPC_GST_FPSCR, mkU64( 1 <<( 31 - crbD ) ),
1ULL << ( 31 - crbD ) );
break;
}
case 0x040: { // mcrfs (Move to Condition Register from FPSCR, PPC32 p465)
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
UChar b21to22 = toUChar( IFIELD( theInstr, 21, 2 ) );
UChar crfS = toUChar( IFIELD( theInstr, 18, 3 ) );
UChar b11to17 = toUChar( IFIELD( theInstr, 11, 7 ) );
IRTemp tmp = newTemp(Ity_I32);
IRExpr* fpscr_all;
if (b21to22 != 0 || b11to17 != 0 || flag_rC != 0) {
vex_printf("dis_fp_scr(ppc)(instr,mcrfs)\n");
return False;
}
DIP("mcrfs crf%d,crf%d\n", crfD, crfS);
vassert(crfD < 8);
vassert(crfS < 8);
fpscr_all = getGST_masked( PPC_GST_FPSCR, MASK_FPSCR_RN );
assign( tmp, binop(Iop_And32,
binop(Iop_Shr32,fpscr_all,mkU8(4 * (7-crfS))),
mkU32(0xF)) );
putGST_field( PPC_GST_CR, mkexpr(tmp), crfD );
break;
}
case 0x046: { // mtfsb0 (Move to FPSCR Bit 0, PPC32 p478)
// Bit crbD of the FPSCR is cleared.
UChar crbD = ifieldRegDS(theInstr);
UInt b11to20 = IFIELD(theInstr, 11, 10);
if (b11to20 != 0) {
vex_printf("dis_fp_scr(ppc)(instr,mtfsb0)\n");
return False;
}
DIP("mtfsb0%s crb%d\n", flag_rC ? ".":"", crbD);
putGST_masked( PPC_GST_FPSCR, mkU64( 0 ), 1ULL << ( 31 - crbD ) );
break;
}
case 0x086: { // mtfsfi (Move to FPSCR Field Immediate, PPC32 p481)
UInt crfD = IFIELD( theInstr, 23, 3 );
UChar b16to22 = toUChar( IFIELD( theInstr, 16, 7 ) );
UChar IMM = toUChar( IFIELD( theInstr, 12, 4 ) );
UChar b11 = toUChar( IFIELD( theInstr, 11, 1 ) );
UChar Wbit;
if (b16to22 != 0 || b11 != 0) {
vex_printf("dis_fp_scr(ppc)(instr,mtfsfi)\n");
return False;
}
DIP("mtfsfi%s crf%d,%d\n", flag_rC ? ".":"", crfD, IMM);
if (GX_level) {
/* This implies that Decimal Floating Point is supported, and the
* FPSCR must be managed as a 64-bit register.
*/
Wbit = toUChar( IFIELD(theInstr, 16, 1) );
} else {
Wbit = 0;
}
crfD = crfD + (8 * (1 - Wbit) );
putGST_field( PPC_GST_FPSCR, mkU32( IMM ), crfD );
break;
}
case 0x247: { // mffs (Move from FPSCR, PPC32 p468)
UChar frD_addr = ifieldRegDS(theInstr);
UInt b11to20 = IFIELD(theInstr, 11, 10);
IRExpr* fpscr_lower = getGST_masked( PPC_GST_FPSCR, MASK_FPSCR_RN );
IRExpr* fpscr_upper = getGST_masked_upper( PPC_GST_FPSCR,
MASK_FPSCR_DRN );
if (b11to20 != 0) {
vex_printf("dis_fp_scr(ppc)(instr,mffs)\n");
return False;
}
DIP("mffs%s fr%u\n", flag_rC ? ".":"", frD_addr);
putFReg( frD_addr,
unop( Iop_ReinterpI64asF64,
binop( Iop_32HLto64, fpscr_upper, fpscr_lower ) ) );
break;
}
case 0x2C7: { // mtfsf (Move to FPSCR Fields, PPC32 p480)
UChar b25 = toUChar( IFIELD(theInstr, 25, 1) );
UChar FM = toUChar( IFIELD(theInstr, 17, 8) );
UChar frB_addr = ifieldRegB(theInstr);
IRTemp frB = newTemp(Ity_F64);
IRTemp rB_64 = newTemp( Ity_I64 );
Int i;
ULong mask;
UChar Wbit;
#define BFP_MASK_SEED 0x3000000000000000ULL
#define DFP_MASK_SEED 0x7000000000000000ULL
if (GX_level) {
/* This implies that Decimal Floating Point is supported, and the
* FPSCR must be managed as a 64-bit register.
*/
Wbit = toUChar( IFIELD(theInstr, 16, 1) );
} else {
Wbit = 0;
}
if (b25 == 1) {
/* new 64 bit move variant for power 6. If L field (bit 25) is
* a one do a full 64 bit move. Note, the FPSCR is not really
* properly modeled. This instruciton only changes the value of
* the rounding mode. The HW exception bits do not get set in
* the simulator. 1/12/09
*/
DIP("mtfsf%s %d,fr%u (L=1)\n", flag_rC ? ".":"", FM, frB_addr);
mask = 0xFF;
} else {
DIP("mtfsf%s %d,fr%u\n", flag_rC ? ".":"", FM, frB_addr);
// Build 32bit mask from FM:
mask = 0;
for (i=0; i<8; i++) {
if ((FM & (1<<(7-i))) == 1) {
/* FPSCR field k is set to the contents of the corresponding
* field of register FRB, where k = i+8x(1-W). In the Power
* ISA, register field numbering is from left to right, so field
* 15 is the least significant field in a 64-bit register. To
* generate the mask, we set all the appropriate rounding mode
* bits in the highest order nibble (field 0) and shift right
* 'k x nibble length'.
*/
if (Wbit)
mask |= DFP_MASK_SEED >> ( 4 * ( i + 8 * ( 1 - Wbit ) ) );
else
mask |= BFP_MASK_SEED >> ( 4 * ( i + 8 * ( 1 - Wbit ) ) );
}
}
}
assign( frB, getFReg(frB_addr));
assign( rB_64, unop( Iop_ReinterpF64asI64, mkexpr( frB ) ) );
putGST_masked( PPC_GST_FPSCR, mkexpr( rB_64 ), mask );
break;
}
default:
vex_printf("dis_fp_scr(ppc)(opc2)\n");
return False;
}
return True;
}
/*------------------------------------------------------------*/
/*--- Decimal Floating Point (DFP) Helper functions ---*/
/*------------------------------------------------------------*/
#define DFP_LONG 1
#define DFP_EXTND 2
#define DFP_LONG_BIAS 398
#define DFP_LONG_ENCODED_FIELD_MASK 0x1F00
#define DFP_EXTND_BIAS 6176
#define DFP_EXTND_ENCODED_FIELD_MASK 0x1F000
#define DFP_LONG_EXP_MSK 0XFF
#define DFP_EXTND_EXP_MSK 0XFFF
#define DFP_G_FIELD_LONG_MASK 0x7FFC0000 // upper 32-bits only
#define DFP_LONG_GFIELD_RT_SHIFT (63 - 13 - 32) // adj for upper 32-bits
#define DFP_G_FIELD_EXTND_MASK 0x7FFFC000 // upper 32-bits only
#define DFP_EXTND_GFIELD_RT_SHIFT (63 - 17 - 32) //adj for upper 32 bits
#define DFP_T_FIELD_LONG_MASK 0x3FFFF // mask for upper 32-bits
#define DFP_T_FIELD_EXTND_MASK 0x03FFFF // mask for upper 32-bits
#define DFP_LONG_EXP_MAX 369 // biased max
#define DFP_LONG_EXP_MIN 0 // biased min
#define DFP_EXTND_EXP_MAX 6111 // biased max
#define DFP_EXTND_EXP_MIN 0 // biased min
#define DFP_LONG_MAX_SIG_DIGITS 16
#define DFP_EXTND_MAX_SIG_DIGITS 34
#define MAX_DIGITS_IN_STRING 8
#define AND(x, y) binop( Iop_And32, x, y )
#define AND4(w, x, y, z) AND( AND( w, x ), AND( y, z ) )
#define OR(x, y) binop( Iop_Or32, x, y )
#define OR3(x, y, z) OR( x, OR( y, z ) )
#define OR4(w, x, y, z) OR( OR( w, x ), OR( y, z ) )
#define NOT(x) unop( Iop_1Uto32, unop( Iop_Not1, unop( Iop_32to1, mkexpr( x ) ) ) )
#define SHL(value, by) binop( Iop_Shl32, value, mkU8( by ) )
#define SHR(value, by) binop( Iop_Shr32, value, mkU8( by ) )
#define BITS5(_b4,_b3,_b2,_b1,_b0) \
(((_b4) << 4) | ((_b3) << 3) | ((_b2) << 2) | \
((_b1) << 1) | ((_b0) << 0))
static IRExpr * Gfield_encoding( IRExpr * lmexp, IRExpr * lmd32 )
{
IRTemp lmd_07_mask = newTemp( Ity_I32 );
IRTemp lmd_8_mask = newTemp( Ity_I32 );
IRTemp lmd_9_mask = newTemp( Ity_I32 );
IRTemp lmexp_00_mask = newTemp( Ity_I32 );
IRTemp lmexp_01_mask = newTemp( Ity_I32 );
IRTemp lmexp_10_mask = newTemp( Ity_I32 );
IRTemp lmd_07_val = newTemp( Ity_I32 );
IRTemp lmd_8_val = newTemp( Ity_I32 );
IRTemp lmd_9_val = newTemp( Ity_I32 );
/* The encodig is as follows:
* lmd - left most digit
* lme - left most 2-bits of the exponent
*
* lmd
* 0 - 7 (lmexp << 3) | lmd
* 8 0b11000 (24 decimal) if lme=0b00;
* 0b11010 (26 decimal) if lme=0b01;
* 0b11100 (28 decimal) if lme=0b10;
* 9 0b11001 (25 decimal) if lme=0b00;
* 0b11011 (27 decimal) if lme=0b01;
* 0b11101 (29 decimal) if lme=0b10;
*/
/* Generate the masks for each condition */
assign( lmd_07_mask,
unop( Iop_1Sto32, binop( Iop_CmpLE32U, lmd32, mkU32( 7 ) ) ) );
assign( lmd_8_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, lmd32, mkU32( 8 ) ) ) );
assign( lmd_9_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, lmd32, mkU32( 9 ) ) ) );
assign( lmexp_00_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, lmexp, mkU32( 0 ) ) ) );
assign( lmexp_01_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, lmexp, mkU32( 1 ) ) ) );
assign( lmexp_10_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, lmexp, mkU32( 2 ) ) ) );
/* Generate the values for each LMD condition, assuming the condition
* is TRUE.
*/
assign( lmd_07_val,
binop( Iop_Or32, binop( Iop_Shl32, lmexp, mkU8( 3 ) ), lmd32 ) );
assign( lmd_8_val,
binop( Iop_Or32,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( lmexp_00_mask ),
mkU32( 24 ) ),
binop( Iop_And32,
mkexpr( lmexp_01_mask ),
mkU32( 26 ) ) ),
binop( Iop_And32, mkexpr( lmexp_10_mask ), mkU32( 28 ) ) ) );
assign( lmd_9_val,
binop( Iop_Or32,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( lmexp_00_mask ),
mkU32( 25 ) ),
binop( Iop_And32,
mkexpr( lmexp_01_mask ),
mkU32( 27 ) ) ),
binop( Iop_And32, mkexpr( lmexp_10_mask ), mkU32( 29 ) ) ) );
/* generate the result from the possible LMD values */
return binop( Iop_Or32,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( lmd_07_mask ),
mkexpr( lmd_07_val ) ),
binop( Iop_And32,
mkexpr( lmd_8_mask ),
mkexpr( lmd_8_val ) ) ),
binop( Iop_And32, mkexpr( lmd_9_mask ), mkexpr( lmd_9_val ) ) );
}
static void Get_lmd( IRTemp * lmd, IRExpr * gfield_0_4 )
{
/* Extract the exponent and the left most digit of the mantissa
* from the G field bits [0:4].
*/
IRTemp lmd_07_mask = newTemp( Ity_I32 );
IRTemp lmd_8_00_mask = newTemp( Ity_I32 );
IRTemp lmd_8_01_mask = newTemp( Ity_I32 );
IRTemp lmd_8_10_mask = newTemp( Ity_I32 );
IRTemp lmd_9_00_mask = newTemp( Ity_I32 );
IRTemp lmd_9_01_mask = newTemp( Ity_I32 );
IRTemp lmd_9_10_mask = newTemp( Ity_I32 );
IRTemp lmd_07_val = newTemp( Ity_I32 );
IRTemp lmd_8_val = newTemp( Ity_I32 );
IRTemp lmd_9_val = newTemp( Ity_I32 );
/* The left most digit (LMD) encoding is as follows:
* lmd
* 0 - 7 (lmexp << 3) | lmd
* 8 0b11000 (24 decimal) if lme=0b00;
* 0b11010 (26 decimal) if lme=0b01;
* 0b11100 (28 decimal) if lme=0b10
* 9 0b11001 (25 decimal) if lme=0b00;
* 0b11011 (27 decimal) if lme=0b01;
* 0b11101 (29 decimal) if lme=0b10;
*/
/* Generate the masks for each condition of LMD and exponent bits */
assign( lmd_07_mask,
unop( Iop_1Sto32, binop( Iop_CmpLE32U,
gfield_0_4,
mkU32( BITS5(1,0,1,1,1) ) ) ) );
assign( lmd_8_00_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,0,0,0) ) ) ) );
assign( lmd_8_01_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,0,1,0) ) ) ) );
assign( lmd_8_10_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,1,0,0) ) ) ) );
assign( lmd_9_00_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,0,0,1) ) ) ) );
assign( lmd_9_01_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,0,1,1) ) ) ) );
assign( lmd_9_10_mask,
unop( Iop_1Sto32, binop( Iop_CmpEQ32,
gfield_0_4,
mkU32( BITS5(1,1,1,0,1) ) ) ) );
/* Generate the values for each LMD condition, assuming the condition
* is TRUE.
*/
assign( lmd_07_val, binop( Iop_And32, gfield_0_4, mkU32( 0x7 ) ) );
assign( lmd_8_val, mkU32( 0x8 ) );
assign( lmd_9_val, mkU32( 0x9 ) );
assign( *lmd,
OR( OR3 ( AND( mkexpr( lmd_07_mask ), mkexpr( lmd_07_val ) ),
AND( mkexpr( lmd_8_00_mask ), mkexpr( lmd_8_val ) ),
AND( mkexpr( lmd_8_01_mask ), mkexpr( lmd_8_val ) )),
OR4( AND( mkexpr( lmd_8_10_mask ), mkexpr( lmd_8_val ) ),
AND( mkexpr( lmd_9_00_mask ), mkexpr( lmd_9_val ) ),
AND( mkexpr( lmd_9_01_mask ), mkexpr( lmd_9_val ) ),
AND( mkexpr( lmd_9_10_mask ), mkexpr( lmd_9_val ) )
) ) );
}
#define DIGIT1_SHR 4 // shift digit 1 to bottom 4 bits
#define DIGIT2_SHR 8 // shift digit 2 to bottom 4 bits
#define DIGIT3_SHR 12
#define DIGIT4_SHR 16
#define DIGIT5_SHR 20
#define DIGIT6_SHR 24
#define DIGIT7_SHR 28
static IRExpr * bcd_digit_inval( IRExpr * bcd_u, IRExpr * bcd_l )
{
/* 60-bit BCD string stored in two 32-bit values. Check that each,
* digit is a valid BCD number, i.e. less then 9.
*/
IRTemp valid = newTemp( Ity_I32 );
assign( valid,
AND4( AND4 ( unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
bcd_l,
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT1_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT2_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT3_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ) ),
AND4 ( unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT4_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT5_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT6_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_l,
mkU8 ( DIGIT7_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ) ),
AND4( unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
bcd_u,
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT1_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT2_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT3_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ) ),
AND4( unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT4_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT5_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT6_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpLE32U,
binop( Iop_And32,
binop( Iop_Shr32,
bcd_u,
mkU8 ( DIGIT7_SHR ) ),
mkU32 ( 0xF ) ),
mkU32( 0x9 ) ) ) ) ) );
return unop( Iop_Not32, mkexpr( valid ) );
}
#undef DIGIT1_SHR
#undef DIGIT2_SHR
#undef DIGIT3_SHR
#undef DIGIT4_SHR
#undef DIGIT5_SHR
#undef DIGIT6_SHR
#undef DIGIT7_SHR
static IRExpr * Generate_neg_sign_mask( IRExpr * sign )
{
return binop( Iop_Or32,
unop( Iop_1Sto32, binop( Iop_CmpEQ32, sign, mkU32( 0xB ) ) ),
unop( Iop_1Sto32, binop( Iop_CmpEQ32, sign, mkU32( 0xD ) ) )
);
}
static IRExpr * Generate_pos_sign_mask( IRExpr * sign )
{
return binop( Iop_Or32,
binop( Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32, sign, mkU32( 0xA ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32, sign, mkU32( 0xC ) ) ) ),
binop( Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32, sign, mkU32( 0xE ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32, sign, mkU32( 0xF ) ) ) ) );
}
static IRExpr * Generate_sign_bit( IRExpr * pos_sign_mask,
IRExpr * neg_sign_mask )
{
return binop( Iop_Or32,
binop( Iop_And32, neg_sign_mask, mkU32( 0x80000000 ) ),
binop( Iop_And32, pos_sign_mask, mkU32( 0x00000000 ) ) );
}
static IRExpr * Generate_inv_mask( IRExpr * invalid_bcd_mask,
IRExpr * pos_sign_mask,
IRExpr * neg_sign_mask )
/* first argument is all 1's if the BCD string had an invalid digit in it. */
{
return binop( Iop_Or32,
invalid_bcd_mask,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
binop( Iop_Or32, pos_sign_mask, neg_sign_mask ),
mkU32( 0x0 ) ) ) );
}
static void Generate_132_bit_bcd_string( IRExpr * frBI64_hi, IRExpr * frBI64_lo,
IRTemp * top_12_l, IRTemp * mid_60_u,
IRTemp * mid_60_l, IRTemp * low_60_u,
IRTemp * low_60_l)
{
IRTemp tmplow60 = newTemp( Ity_I64 );
IRTemp tmpmid60 = newTemp( Ity_I64 );
IRTemp tmptop12 = newTemp( Ity_I64 );
IRTemp low_50 = newTemp( Ity_I64 );
IRTemp mid_50 = newTemp( Ity_I64 );
IRTemp top_10 = newTemp( Ity_I64 );
IRTemp top_12_u = newTemp( Ity_I32 ); // only needed for a dummy arg
/* Convert the 110-bit densely packed BCD string to a 128-bit BCD string */
/* low_50[49:0] = ((frBI64_lo[49:32] << 14) | frBI64_lo[31:0]) */
assign( low_50,
binop( Iop_32HLto64,
binop( Iop_And32,
unop( Iop_64HIto32, frBI64_lo ),
mkU32( 0x3FFFF ) ),
unop( Iop_64to32, frBI64_lo ) ) );
/* Convert the 50 bit densely packed BCD string to a 60 bit
* BCD string.
*/
assign( tmplow60, unop( Iop_DPBtoBCD, mkexpr( low_50 ) ) );
assign( *low_60_u, unop( Iop_64HIto32, mkexpr( tmplow60 ) ) );
assign( *low_60_l, unop( Iop_64to32, mkexpr( tmplow60 ) ) );
/* mid_50[49:0] = ((frBI64_hi[35:32] << 14) | frBI64_hi[31:18]) |
* ((frBI64_hi[17:0] << 14) | frBI64_lo[63:50])
*/
assign( mid_50,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32,
binop( Iop_And32,
unop( Iop_64HIto32, frBI64_hi ),
mkU32( 0xF ) ),
mkU8( 14 ) ),
binop( Iop_Shr32,
unop( Iop_64to32, frBI64_hi ),
mkU8( 18 ) ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
unop( Iop_64to32, frBI64_hi ),
mkU8( 14 ) ),
binop( Iop_Shr32,
unop( Iop_64HIto32, frBI64_lo ),
mkU8( 18 ) ) ) ) );
/* Convert the 50 bit densely packed BCD string to a 60 bit
* BCD string.
*/
assign( tmpmid60, unop( Iop_DPBtoBCD, mkexpr( mid_50 ) ) );
assign( *mid_60_u, unop( Iop_64HIto32, mkexpr( tmpmid60 ) ) );
assign( *mid_60_l, unop( Iop_64to32, mkexpr( tmpmid60 ) ) );
/* top_10[49:0] = frBI64_hi[45:36]) | */
assign( top_10,
binop( Iop_32HLto64,
mkU32( 0 ),
binop( Iop_And32,
binop( Iop_Shr32,
unop( Iop_64HIto32, frBI64_hi ),
mkU8( 4 ) ),
mkU32( 0x3FF ) ) ) );
/* Convert the 10 bit densely packed BCD string to a 12 bit
* BCD string.
*/
assign( tmptop12, unop( Iop_DPBtoBCD, mkexpr( top_10 ) ) );
assign( top_12_u, unop( Iop_64HIto32, mkexpr( tmptop12 ) ) );
assign( *top_12_l, unop( Iop_64to32, mkexpr( tmptop12 ) ) );
}
static void Count_zeros( int start, IRExpr * init_cnt, IRExpr * init_flag,
IRTemp * final_cnt, IRTemp * final_flag,
IRExpr * string )
{
IRTemp cnt[MAX_DIGITS_IN_STRING + 1];IRTemp flag[MAX_DIGITS_IN_STRING+1];
int digits = MAX_DIGITS_IN_STRING;
int i;
cnt[start-1] = newTemp( Ity_I8 );
flag[start-1] = newTemp( Ity_I8 );
assign( cnt[start-1], init_cnt);
assign( flag[start-1], init_flag);
for ( i = start; i <= digits; i++) {
cnt[i] = newTemp( Ity_I8 );
flag[i] = newTemp( Ity_I8 );
assign( cnt[i],
binop( Iop_Add8,
mkexpr( cnt[i-1] ),
binop(Iop_And8,
unop( Iop_1Uto8,
binop(Iop_CmpEQ32,
binop(Iop_And32,
string,
mkU32( 0xF <<
( ( digits - i ) * 4) ) ),
mkU32( 0 ) ) ),
binop( Iop_Xor8, /* complement flag */
mkexpr( flag[i - 1] ),
mkU8( 0xFF ) ) ) ) );
/* set flag to 1 if digit was not a zero */
assign( flag[i],
binop(Iop_Or8,
unop( Iop_1Sto8,
binop(Iop_CmpNE32,
binop(Iop_And32,
string,
mkU32( 0xF <<
( (digits - i) * 4) ) ),
mkU32( 0 ) ) ),
mkexpr( flag[i - 1] ) ) );
}
*final_cnt = cnt[digits];
*final_flag = flag[digits];
}
static IRExpr * Count_leading_zeros_60( IRExpr * lmd, IRExpr * upper_28,
IRExpr * low_32 )
{
IRTemp num_lmd = newTemp( Ity_I8 );
IRTemp num_upper = newTemp( Ity_I8 );
IRTemp num_low = newTemp( Ity_I8 );
IRTemp lmd_flag = newTemp( Ity_I8 );
IRTemp upper_flag = newTemp( Ity_I8 );
IRTemp low_flag = newTemp( Ity_I8 );
assign( num_lmd, unop( Iop_1Uto8, binop( Iop_CmpEQ32, lmd, mkU32( 0 ) ) ) );
assign( lmd_flag, unop( Iop_Not8, mkexpr( num_lmd ) ) );
Count_zeros( 2,
mkexpr( num_lmd ),
mkexpr( lmd_flag ),
&num_upper,
&upper_flag,
upper_28 );
Count_zeros( 1,
mkexpr( num_upper ),
mkexpr( upper_flag ),
&num_low,
&low_flag,
low_32 );
return mkexpr( num_low );
}
static IRExpr * Count_leading_zeros_128( IRExpr * lmd, IRExpr * top_12_l,
IRExpr * mid_60_u, IRExpr * mid_60_l,
IRExpr * low_60_u, IRExpr * low_60_l)
{
IRTemp num_lmd = newTemp( Ity_I8 );
IRTemp num_top = newTemp( Ity_I8 );
IRTemp num_mid_u = newTemp( Ity_I8 );
IRTemp num_mid_l = newTemp( Ity_I8 );
IRTemp num_low_u = newTemp( Ity_I8 );
IRTemp num_low_l = newTemp( Ity_I8 );
IRTemp lmd_flag = newTemp( Ity_I8 );
IRTemp top_flag = newTemp( Ity_I8 );
IRTemp mid_u_flag = newTemp( Ity_I8 );
IRTemp mid_l_flag = newTemp( Ity_I8 );
IRTemp low_u_flag = newTemp( Ity_I8 );
IRTemp low_l_flag = newTemp( Ity_I8 );
/* Check the LMD, digit 16, to see if it is zero. */
assign( num_lmd, unop( Iop_1Uto8, binop( Iop_CmpEQ32, lmd, mkU32( 0 ) ) ) );
assign( lmd_flag, unop( Iop_Not8, mkexpr( num_lmd ) ) );
Count_zeros( 6,
mkexpr( num_lmd ),
mkexpr( lmd_flag ),
&num_top,
&top_flag,
top_12_l );
Count_zeros( 1,
mkexpr( num_top ),
mkexpr( top_flag ),
&num_mid_u,
&mid_u_flag,
binop( Iop_Or32,
binop( Iop_Shl32, mid_60_u, mkU8( 2 ) ),
binop( Iop_Shr32, mid_60_l, mkU8( 30 ) ) ) );
Count_zeros( 2,
mkexpr( num_mid_u ),
mkexpr( mid_u_flag ),
&num_mid_l,
&mid_l_flag,
mid_60_l );
Count_zeros( 1,
mkexpr( num_mid_l ),
mkexpr( mid_l_flag ),
&num_low_u,
&low_u_flag,
binop( Iop_Or32,
binop( Iop_Shl32, low_60_u, mkU8( 2 ) ),
binop( Iop_Shr32, low_60_l, mkU8( 30 ) ) ) );
Count_zeros( 2,
mkexpr( num_low_u ),
mkexpr( low_u_flag ),
&num_low_l,
&low_l_flag,
low_60_l );
return mkexpr( num_low_l );
}
static IRExpr * Check_unordered(IRExpr * val)
{
IRTemp gfield0to5 = newTemp( Ity_I32 );
/* Extract G[0:4] */
assign( gfield0to5,
binop( Iop_And32,
binop( Iop_Shr32, unop( Iop_64HIto32, val ), mkU8( 26 ) ),
mkU32( 0x1F ) ) );
/* Check for unordered, return all 1'x if true */
return binop( Iop_Or32, /* QNaN check */
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield0to5 ),
mkU32( 0x1E ) ) ),
unop( Iop_1Sto32, /* SNaN check */
binop( Iop_CmpEQ32,
mkexpr( gfield0to5 ),
mkU32( 0x1F ) ) ) );
}
#undef AND
#undef AND4
#undef OR
#undef OR3
#undef OR4
#undef NOT
#undef SHR
#undef SHL
#undef BITS5
/*------------------------------------------------------------*/
/*--- Decimal Floating Point (DFP) instruction translation ---*/
/*------------------------------------------------------------*/
/* DFP Arithmetic instructions */
static Bool dis_dfp_arith(UInt theInstr)
{
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frA = newTemp( Ity_D64 );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D64 );
IRExpr* round = get_IR_roundingmode_DFP();
/* By default, if flag_RC is set, we will clear cr1 after the
* operation. In reality we should set cr1 to indicate the
* exception status of the operation, but since we're not
* simulating exceptions, the exception status will appear to be
* zero. Hence cr1 should be cleared if this is a . form insn.
*/
Bool clear_CR1 = True;
assign( frA, getDReg( frA_addr ) );
assign( frB, getDReg( frB_addr ) );
switch (opc2) {
case 0x2: // dadd
DIP( "dadd%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_AddD64, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x202: // dsub
DIP( "dsub%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_SubD64, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x22: // dmul
DIP( "dmul%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_MulD64, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x222: // ddiv
DIP( "ddiv%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_DivD64, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
}
putDReg( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* Quad DFP Arithmetic instructions */
static Bool dis_dfp_arithq(UInt theInstr)
{
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frA = newTemp( Ity_D128 );
IRTemp frB = newTemp( Ity_D128 );
IRTemp frS = newTemp( Ity_D128 );
IRExpr* round = get_IR_roundingmode_DFP();
/* By default, if flag_RC is set, we will clear cr1 after the
* operation. In reality we should set cr1 to indicate the
* exception status of the operation, but since we're not
* simulating exceptions, the exception status will appear to be
* zero. Hence cr1 should be cleared if this is a . form insn.
*/
Bool clear_CR1 = True;
assign( frA, getDReg_pair( frA_addr ) );
assign( frB, getDReg_pair( frB_addr ) );
switch (opc2) {
case 0x2: // daddq
DIP( "daddq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_AddD128, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x202: // dsubq
DIP( "dsubq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_SubD128, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x22: // dmulq
DIP( "dmulq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_MulD128, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x222: // ddivq
DIP( "ddivq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, triop( Iop_DivD128, round, mkexpr( frA ), mkexpr( frB ) ) );
break;
}
putDReg_pair( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* DFP 64-bit logical shift instructions */
static Bool dis_dfp_shift(UInt theInstr) {
UInt opc2 = ifieldOPClo9( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar shift_val = IFIELD(theInstr, 10, 6);
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frA = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D64 );
Bool clear_CR1 = True;
assign( frA, getDReg( frA_addr ) );
switch (opc2) {
case 0x42: // dscli
DIP( "dscli%s fr%u,fr%u,%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, shift_val );
assign( frS, binop( Iop_ShlD64, mkexpr( frA ), mkU8( shift_val ) ) );
break;
case 0x62: // dscri
DIP( "dscri%s fr%u,fr%u,%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, shift_val );
assign( frS, binop( Iop_ShrD64, mkexpr( frA ), mkU8( shift_val ) ) );
break;
}
putDReg( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* Quad DFP logical shift instructions */
static Bool dis_dfp_shiftq(UInt theInstr) {
UInt opc2 = ifieldOPClo9( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar shift_val = IFIELD(theInstr, 10, 6);
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frA = newTemp( Ity_D128 );
IRTemp frS = newTemp( Ity_D128 );
Bool clear_CR1 = True;
assign( frA, getDReg_pair( frA_addr ) );
switch (opc2) {
case 0x42: // dscliq
DIP( "dscliq%s fr%u,fr%u,%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, shift_val );
assign( frS, binop( Iop_ShlD128, mkexpr( frA ), mkU8( shift_val ) ) );
break;
case 0x62: // dscriq
DIP( "dscriq%s fr%u,fr%u,%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, shift_val );
assign( frS, binop( Iop_ShrD128, mkexpr( frA ), mkU8( shift_val ) ) );
break;
}
putDReg_pair( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* DFP 64-bit format conversion instructions */
static Bool dis_dfp_fmt_conv(UInt theInstr) {
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
IRExpr* round = get_IR_roundingmode_DFP();
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frB;
IRTemp frS;
Bool clear_CR1 = True;
switch (opc2) {
case 0x102: //dctdp
DIP( "dctdp%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
frB = newTemp( Ity_D64 );
frS = newTemp( Ity_D64 );
assign( frB, getDReg( frB_addr ) );
assign( frS, unop( Iop_D32toD64, mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS ) );
break;
case 0x302: // drsp
DIP( "drsp%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
frB = newTemp( Ity_D64 );
frS = newTemp( Ity_D64 );
assign( frB, getDReg( frB_addr ) );
assign( frS, binop( Iop_D64toD32, round, mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS ) );
break;
case 0x122: // dctfix
DIP( "dctfix%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
frB = newTemp( Ity_D64 );
frS = newTemp( Ity_D64 );
assign( frB, getDReg( frB_addr ) );
assign( frS, binop( Iop_D64toI64S, round, mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS ) );
break;
case 0x322: // dcffix
DIP( "dcffix%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
frB = newTemp( Ity_D64 );
frS = newTemp( Ity_D64 );
assign( frB, getDReg( frB_addr ) );
assign( frS, binop( Iop_I64StoD64, round, mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS ) );
break;
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* Quad DFP format conversion instructions */
static Bool dis_dfp_fmt_convq(UInt theInstr) {
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
IRExpr* round = get_IR_roundingmode_DFP();
IRTemp frB64 = newTemp( Ity_D64 );
IRTemp frB128 = newTemp( Ity_D128 );
IRTemp frS64 = newTemp( Ity_D64 );
IRTemp frS128 = newTemp( Ity_D128 );
UChar flag_rC = ifieldBIT0( theInstr );
Bool clear_CR1 = True;
switch (opc2) {
case 0x102: // dctqpq
DIP( "dctqpq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
assign( frB64, getDReg( frB_addr ) );
assign( frS128, unop( Iop_D64toD128, mkexpr( frB64 ) ) );
putDReg_pair( frS_addr, mkexpr( frS128 ) );
break;
case 0x122: // dctfixq
DIP( "dctfixq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
assign( frB128, getDReg_pair( frB_addr ) );
assign( frS64, binop( Iop_D128toI64S, round, mkexpr( frB128 ) ) );
putDReg( frS_addr, mkexpr( frS64 ) );
break;
case 0x302: //drdpq
DIP( "drdpq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
assign( frB128, getDReg_pair( frB_addr ) );
assign( frS64, binop( Iop_D128toD64, round, mkexpr( frB128 ) ) );
putDReg( frS_addr, mkexpr( frS64 ) );
break;
case 0x322: // dcffixq
/* Have to introduce an IOP for this instruction so it will work
* on POWER 6 because emulating the instruction requires a POWER 7
* DFP instruction in the emulation code.
*/
DIP( "dcffixq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
assign( frB64, getDReg( frB_addr ) );
assign( frS128, unop( Iop_I64StoD128, mkexpr( frB64 ) ) );
putDReg_pair( frS_addr, mkexpr( frS128 ) );
break;
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_round( UInt theInstr ) {
UChar frS_addr = ifieldRegDS(theInstr);
UChar R = IFIELD(theInstr, 16, 1);
UChar RMC = IFIELD(theInstr, 9, 2);
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D64 );
UInt opc2 = ifieldOPClo8( theInstr );
Bool clear_CR1 = True;
switch (opc2) {
/* drintn, is the same as drintx. The only difference is this
* instruction does not generate an exception for an inexact operation.
* Currently not supporting inexact exceptions.
*/
case 0x63: // drintx
case 0xE3: // drintn
DIP( "drintx/drintn%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
/* pass the value of R and RMC in the same field */
assign( frB, getDReg( frB_addr ) );
assign( frS, binop( Iop_RoundD64toInt,
mkU32( ( R << 3 ) | RMC ),
mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS ) );
break;
default:
vex_printf("dis_dfp_round(ppc)(opc2)\n");
return False;
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_roundq(UInt theInstr) {
UChar frS_addr = ifieldRegDS( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar R = IFIELD(theInstr, 16, 1);
UChar RMC = IFIELD(theInstr, 9, 2);
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frB = newTemp( Ity_D128 );
IRTemp frS = newTemp( Ity_D128 );
Bool clear_CR1 = True;
UInt opc2 = ifieldOPClo8( theInstr );
switch (opc2) {
/* drintnq, is the same as drintxq. The only difference is this
* instruction does not generate an exception for an inexact operation.
* Currently not supporting inexact exceptions.
*/
case 0x63: // drintxq
case 0xE3: // drintnq
DIP( "drintxq/drintnq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
/* pass the value of R and RMC in the same field */
assign( frB, getDReg_pair( frB_addr ) );
assign( frS, binop( Iop_RoundD128toInt,
mkU32( ( R << 3 ) | RMC ),
mkexpr( frB ) ) );
putDReg_pair( frS_addr, mkexpr( frS ) );
break;
default:
vex_printf("dis_dfp_roundq(ppc)(opc2)\n");
return False;
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_quantize_sig_rrnd(UInt theInstr) {
UInt opc2 = ifieldOPClo8( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
UInt TE_value = IFIELD(theInstr, 16, 4);
UInt TE_sign = IFIELD(theInstr, 20, 1);
UInt RMC = IFIELD(theInstr, 9, 2);
IRTemp frA = newTemp( Ity_D64 );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D64 );
Bool clear_CR1 = True;
assign( frB, getDReg( frB_addr ) );
switch (opc2) {
case 0x43: // dquai
DIP( "dquai%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
IRTemp TE_D64 = newTemp( Ity_D64 );
/* Generate a reference DFP value frA with the desired exponent
* given by TE using significand from frB. Need to add the bias
* 398 to TE. TE is stored as a 2's complement number.
*/
if (TE_sign == 1) {
/* Take 2's complement of the 5-bit value and subtract from bias.
* Bias is adjusted for the +1 required when taking 2's complement.
*/
assign( TE_D64,
unop( Iop_ReinterpI64asD64,
binop( Iop_Sub64, mkU64( 397 ),
binop( Iop_And64, mkU64( 0xF ),
unop( Iop_Not64, mkU64( TE_value ) )
) ) ) );
} else {
assign( TE_D64,
unop( Iop_ReinterpI64asD64,
binop( Iop_Add64, mkU64( 398 ), mkU64( TE_value ) ) ) );
}
assign( frA, binop( Iop_InsertExpD64, mkexpr( TE_D64 ),
unop( Iop_ReinterpI64asD64, mkU64( 1 ) ) ) );
assign( frS, triop( Iop_QuantizeD64,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
case 0x3: // dqua
DIP( "dqua%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frA, getDReg( frA_addr ) );
assign( frS, triop( Iop_QuantizeD64,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
case 0x23: // drrnd
DIP( "drrnd%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frA, getDReg( frA_addr ) );
assign( frS, triop( Iop_SignificanceRoundD64,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
default:
vex_printf("dis_dfp_quantize_sig_rrnd(ppc)(opc2)\n");
return False;
}
putDReg( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_quantize_sig_rrndq(UInt theInstr) {
UInt opc2 = ifieldOPClo8( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
UInt TE_value = IFIELD(theInstr, 16, 4);
UInt TE_sign = IFIELD(theInstr, 20, 1);
UInt RMC = IFIELD(theInstr, 9, 2);
IRTemp frA = newTemp( Ity_D128 );
IRTemp frB = newTemp( Ity_D128 );
IRTemp frS = newTemp( Ity_D128 );
Bool clear_CR1 = True;
assign( frB, getDReg_pair( frB_addr ) );
switch (opc2) {
case 0x43: // dquaiq
DIP( "dquaiq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
IRTemp TE_D64 = newTemp( Ity_D64 );
/* Generate a reference DFP value frA with the desired exponent
* given by TE using significand of 1. Need to add the bias
* 6176 to TE.
*/
if (TE_sign == 1) {
/* Take 2's complement of the 5-bit value and subtract from bias.
* Bias adjusted for the +1 required when taking 2's complement.
*/
assign( TE_D64,
unop( Iop_ReinterpI64asD64,
binop( Iop_Sub64, mkU64( 6175 ),
binop( Iop_And64, mkU64( 0xF ),
unop( Iop_Not64, mkU64( TE_value ) )
) ) ) );
} else {
assign( TE_D64,
unop( Iop_ReinterpI64asD64,
binop( Iop_Add64, mkU64( 6176 ), mkU64( TE_value ) )
) );
}
assign( frA,
binop( Iop_InsertExpD128, mkexpr( TE_D64 ),
unop( Iop_D64toD128,
unop( Iop_ReinterpI64asD64, mkU64( 1 ) ) ) ) );
assign( frS, triop( Iop_QuantizeD128,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
case 0x3: // dquaq
DIP( "dquaiq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frA, getDReg_pair( frA_addr ) );
assign( frS, triop( Iop_QuantizeD128,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
case 0x23: // drrndq
DIP( "drrndq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frA, getDReg_pair( frA_addr ) );
assign( frS, triop( Iop_SignificanceRoundD128,
mkU32( RMC ),
mkexpr( frA ),
mkexpr( frB ) ) );
break;
default:
vex_printf("dis_dfp_quantize_sig_rrndq(ppc)(opc2)\n");
return False;
}
putDReg_pair( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_extract_insert(UInt theInstr) {
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
Bool clear_CR1 = True;
IRTemp frA = newTemp( Ity_D64 );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D64 );
assign( frA, getDReg( frA_addr ) );
assign( frB, getDReg( frB_addr ) );
switch (opc2) {
case 0x162: // dxex
DIP( "dxex%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, unop( Iop_ExtractExpD64, mkexpr( frB ) ) );
break;
case 0x362: // diex
DIP( "diex%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frS, binop( Iop_InsertExpD64, mkexpr( frA ), mkexpr( frB ) ) );
break;
default:
vex_printf("dis_dfp_extract_insert(ppc)(opc2)\n");
return False;
}
putDReg( frS_addr, mkexpr( frS ) );
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
static Bool dis_dfp_extract_insertq(UInt theInstr) {
UInt opc2 = ifieldOPClo10( theInstr );
UChar frS_addr = ifieldRegDS( theInstr );
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar flag_rC = ifieldBIT0( theInstr );
IRTemp frA = newTemp( Ity_D64 );
IRTemp frB = newTemp( Ity_D128 );
IRTemp frS64 = newTemp( Ity_D64 );
IRTemp frS = newTemp( Ity_D128 );
Bool clear_CR1 = True;
assign( frB, getDReg_pair( frB_addr ) );
switch (opc2) {
case 0x162: // dxexq
DIP( "dxexq%s fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frB_addr );
/* Instruction actually returns a 64-bit result. So as to be
* consistent and not have to add a new struct, the emulation returns
* the 64-bit result in the upper and lower register.
*/
assign( frS64, unop( Iop_ExtractExpD128, mkexpr( frB ) ) );
putDReg( frS_addr, mkexpr( frS64 ) );
break;
case 0x362: // diexq
DIP( "diexq%s fr%u,fr%u,fr%u\n",
flag_rC ? ".":"", frS_addr, frA_addr, frB_addr );
assign( frA, getDReg( frA_addr ) );
assign( frS, binop( Iop_InsertExpD128, mkexpr( frA ), mkexpr( frB ) ) );
putDReg_pair( frS_addr, mkexpr( frS ) );
break;
default:
vex_printf("dis_dfp_extract_insertq(ppc)(opc2)\n");
return False;
}
if (flag_rC && clear_CR1) {
putCR321( 1, mkU8( 0 ) );
putCR0( 1, mkU8( 0 ) );
}
return True;
}
/* DFP 64-bit comparison instructions */
static Bool dis_dfp_compare(UInt theInstr) {
/* X-Form */
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) ); // AKA BF
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UInt opc1 = ifieldOPC( theInstr );
IRTemp frA;
IRTemp frB;
IRTemp ccIR = newTemp( Ity_I32 );
IRTemp ccPPC32 = newTemp( Ity_I32 );
/* Note: Differences between dcmpu and dcmpo are only in exception
flag settings, which aren't supported anyway. */
switch (opc1) {
case 0x3B: /* dcmpo and dcmpu, DFP 64-bit */
DIP( "dcmpo %u,fr%u,fr%u\n", crfD, frA_addr, frB_addr );
frA = newTemp( Ity_D64 );
frB = newTemp( Ity_D64 );
assign( frA, getDReg( frA_addr ) );
assign( frB, getDReg( frB_addr ) );
assign( ccIR, binop( Iop_CmpD64, mkexpr( frA ), mkexpr( frB ) ) );
break;
case 0x3F: /* dcmpoq and dcmpuq,DFP 128-bit */
DIP( "dcmpoq %u,fr%u,fr%u\n", crfD, frA_addr, frB_addr );
frA = newTemp( Ity_D128 );
frB = newTemp( Ity_D128 );
assign( frA, getDReg_pair( frA_addr ) );
assign( frB, getDReg_pair( frB_addr ) );
assign( ccIR, binop( Iop_CmpD128, mkexpr( frA ), mkexpr( frB ) ) );
break;
default:
vex_printf("dis_dfp_compare(ppc)(opc2)\n");
return False;
}
/* Map compare result from IR to PPC32 */
/*
FP cmp result | PPC | IR
--------------------------
UN | 0x1 | 0x45
EQ | 0x2 | 0x40
GT | 0x4 | 0x00
LT | 0x8 | 0x01
*/
assign( ccPPC32,
binop( Iop_Shl32,
mkU32( 1 ),
unop( Iop_32to8,
binop( Iop_Or32,
binop( Iop_And32,
unop( Iop_Not32,
binop( Iop_Shr32,
mkexpr( ccIR ),
mkU8( 5 ) ) ),
mkU32( 2 ) ),
binop( Iop_And32,
binop( Iop_Xor32,
mkexpr( ccIR ),
binop( Iop_Shr32,
mkexpr( ccIR ),
mkU8( 6 ) ) ),
mkU32( 1 ) ) ) ) ) );
putGST_field( PPC_GST_CR, mkexpr( ccPPC32 ), crfD );
return True;
}
/* Test class/group/exponent/significance instructions. */
static Bool dis_dfp_exponent_test ( UInt theInstr )
{
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp frA = newTemp( Ity_D64 );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frA128 = newTemp( Ity_D128 );
IRTemp frB128 = newTemp( Ity_D128 );
UInt opc1 = ifieldOPC( theInstr );
IRTemp gfield_A = newTemp( Ity_I32 );
IRTemp gfield_B = newTemp( Ity_I32 );
IRTemp gfield_mask = newTemp( Ity_I32 );
IRTemp exponent_A = newTemp( Ity_I32 );
IRTemp exponent_B = newTemp( Ity_I32 );
IRTemp A_NaN_true = newTemp( Ity_I32 );
IRTemp B_NaN_true = newTemp( Ity_I32 );
IRTemp A_inf_true = newTemp( Ity_I32 );
IRTemp B_inf_true = newTemp( Ity_I32 );
IRTemp A_equals_B = newTemp( Ity_I32 );
IRTemp finite_number = newTemp( Ity_I32 );
IRTemp cc0 = newTemp( Ity_I32 );
IRTemp cc1 = newTemp( Ity_I32 );
IRTemp cc2 = newTemp( Ity_I32 );
IRTemp cc3 = newTemp( Ity_I32 );
/* The dtstex and dtstexg instructions only differ in the size of the
* exponent field. The following switch statement takes care of the size
* specific setup. Once the value of the exponents, the G-field shift
* and mask is setup the remaining code is identical.
*/
switch (opc1) {
case 0x3b: // dtstex Extended instruction setup
DIP("dtstex %u,r%u,r%d\n", crfD, frA_addr, frB_addr);
assign( frA, getDReg( frA_addr ) );
assign( frB, getDReg( frB_addr ) );
assign( gfield_mask, mkU32( DFP_G_FIELD_LONG_MASK ) );
assign(exponent_A, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD64,
mkexpr( frA ) ) ) ) );
assign(exponent_B, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD64,
mkexpr( frB ) ) ) ) );
break;
case 0x3F: // dtstexq Quad instruction setup
DIP("dtstexq %u,r%u,r%d\n", crfD, frA_addr, frB_addr);
assign( frA128, getDReg_pair( frA_addr ) );
assign( frB128, getDReg_pair( frB_addr ) );
assign( frA, unop( Iop_D128HItoD64, mkexpr( frA128 ) ) );
assign( frB, unop( Iop_D128HItoD64, mkexpr( frB128 ) ) );
assign( gfield_mask, mkU32( DFP_G_FIELD_EXTND_MASK ) );
assign( exponent_A, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD128,
mkexpr( frA128 ) ) ) ) );
assign( exponent_B, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD128,
mkexpr( frB128 ) ) ) ) );
break;
default:
vex_printf("dis_dfp_exponent_test(ppc)(opc2)\n");
return False;
}
/* Extract the Gfield */
assign( gfield_A, binop( Iop_And32,
mkexpr( gfield_mask ),
unop( Iop_64HIto32,
unop( Iop_ReinterpD64asI64,
mkexpr(frA) ) ) ) );
assign( gfield_B, binop( Iop_And32,
mkexpr( gfield_mask ),
unop( Iop_64HIto32,
unop( Iop_ReinterpD64asI64,
mkexpr(frB) ) ) ) );
/* check for NAN */
assign( A_NaN_true, binop(Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_A ),
mkU32( 0x7C000000 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_A ),
mkU32( 0x7E000000 ) )
) ) );
assign( B_NaN_true, binop(Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_B ),
mkU32( 0x7C000000 ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_B ),
mkU32( 0x7E000000 ) )
) ) );
/* check for infinity */
assign( A_inf_true,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_A ),
mkU32( 0x78000000 ) ) ) );
assign( B_inf_true,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( gfield_B ),
mkU32( 0x78000000 ) ) ) );
assign( finite_number,
unop( Iop_Not32,
binop( Iop_Or32,
binop( Iop_Or32,
mkexpr( A_NaN_true ),
mkexpr( B_NaN_true ) ),
binop( Iop_Or32,
mkexpr( A_inf_true ),
mkexpr( B_inf_true ) ) ) ) );
/* Calculate the condition code bits
* If QNaN,SNaN, +infinity, -infinity then cc0, cc1 and cc2 are zero
* regardless of the value of the comparisons and cc3 is 1. Otherwise,
* cc0, cc1 and cc0 reflect the results of the comparisons.
*/
assign( A_equals_B,
binop( Iop_Or32,
unop( Iop_1Uto32,
binop( Iop_CmpEQ32,
mkexpr( exponent_A ),
mkexpr( exponent_B ) ) ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( A_inf_true ),
mkexpr( B_inf_true ) ),
binop( Iop_And32,
mkexpr( A_NaN_true ),
mkexpr( B_NaN_true ) ) ) ) );
assign( cc0, binop( Iop_And32,
mkexpr( finite_number ),
binop( Iop_Shl32,
unop( Iop_1Uto32,
binop( Iop_CmpLT32U,
mkexpr( exponent_A ),
mkexpr( exponent_B ) ) ),
mkU8( 3 ) ) ) );
assign( cc1, binop( Iop_And32,
mkexpr( finite_number ),
binop( Iop_Shl32,
unop( Iop_1Uto32,
binop( Iop_CmpLT32U,
mkexpr( exponent_B ),
mkexpr( exponent_A ) ) ),
mkU8( 2 ) ) ) );
assign( cc2, binop( Iop_Shl32,
binop( Iop_And32,
mkexpr( A_equals_B ),
mkU32( 1 ) ),
mkU8( 1 ) ) );
assign( cc3, binop( Iop_And32,
unop( Iop_Not32, mkexpr( A_equals_B ) ),
binop( Iop_And32,
mkU32( 0x1 ),
binop( Iop_Or32,
binop( Iop_Or32,
mkexpr ( A_inf_true ),
mkexpr ( B_inf_true ) ),
binop( Iop_Or32,
mkexpr ( A_NaN_true ),
mkexpr ( B_NaN_true ) ) )
) ) );
/* store the condition code */
putGST_field( PPC_GST_CR,
binop( Iop_Or32,
mkexpr( cc0 ),
binop( Iop_Or32,
mkexpr( cc1 ),
binop( Iop_Or32,
mkexpr( cc2 ),
mkexpr( cc3 ) ) ) ),
crfD );
return True;
}
/* Test class/group/exponent/significance instructions. */
static Bool dis_dfp_class_test ( UInt theInstr )
{
UChar frA_addr = ifieldRegA( theInstr );
IRTemp frA = newTemp( Ity_D64 );
IRTemp abs_frA = newTemp( Ity_D64 );
IRTemp frAI64_hi = newTemp( Ity_I64 );
IRTemp frAI64_lo = newTemp( Ity_I64 );
UInt opc1 = ifieldOPC( theInstr );
UInt opc2 = ifieldOPClo9( theInstr );
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) ); // AKA BF
UInt DCM = IFIELD( theInstr, 10, 6 );
IRTemp DCM_calc = newTemp( Ity_I32 );
UInt max_exp = 0;
UInt min_exp = 0;
IRTemp min_subnormalD64 = newTemp( Ity_D64 );
IRTemp min_subnormalD128 = newTemp( Ity_D128 );
IRTemp significand64 = newTemp( Ity_D64 );
IRTemp significand128 = newTemp( Ity_D128 );
IRTemp exp_min_normal = newTemp( Ity_D64 );
IRTemp exponent = newTemp( Ity_I32 );
IRTemp infinity_true = newTemp( Ity_I32 );
IRTemp SNaN_true = newTemp( Ity_I32 );
IRTemp QNaN_true = newTemp( Ity_I32 );
IRTemp subnormal_true = newTemp( Ity_I32 );
IRTemp normal_true = newTemp( Ity_I32 );
IRTemp extreme_true = newTemp( Ity_I32 );
IRTemp lmd = newTemp( Ity_I32 );
IRTemp lmd_zero_true = newTemp( Ity_I32 );
IRTemp zero_true = newTemp( Ity_I32 );
IRTemp sign = newTemp( Ity_I32 );
IRTemp field = newTemp( Ity_I32 );
IRTemp ccIR_zero = newTemp( Ity_I32 );
IRTemp ccIR_subnormal = newTemp( Ity_I32 );
/* UInt size = DFP_LONG; JRS:unused */
IRTemp gfield = newTemp( Ity_I32 );
IRTemp gfield_0_4_shift = newTemp( Ity_I8 );
IRTemp gfield_mask = newTemp( Ity_I32 );
IRTemp dcm0 = newTemp( Ity_I32 );
IRTemp dcm1 = newTemp( Ity_I32 );
IRTemp dcm2 = newTemp( Ity_I32 );
IRTemp dcm3 = newTemp( Ity_I32 );
IRTemp dcm4 = newTemp( Ity_I32 );
IRTemp dcm5 = newTemp( Ity_I32 );
/* The only difference between the dtstdc and dtstdcq instructions is
* size of the T and G fields. The calculation of the 4 bit field
* is the same. Setup the parameters and values that are DFP size
* specific. The rest of the code is independent of the DFP size.
*
* The Io_CmpD64 is used below. The instruction sets the ccIR values.
* The interpretation of the ccIR values is as follows:
*
* DFP cmp result | IR
* --------------------------
* UN | 0x45
* EQ | 0x40
* GT | 0x00
* LT | 0x01
*/
assign( frA, getDReg( frA_addr ) );
assign( frAI64_hi, unop( Iop_ReinterpD64asI64, mkexpr( frA ) ) );
assign( abs_frA, unop( Iop_ReinterpI64asD64,
binop( Iop_And64,
unop( Iop_ReinterpD64asI64,
mkexpr( frA ) ),
mkU64( 0x7FFFFFFFFFFFFFFFULL ) ) ) );
assign( gfield_0_4_shift, mkU8( 31 - 5 ) ); // G-field[0:4]
switch (opc1) {
case 0x3b: // dtstdc, dtstdg
DIP("dtstd%s %u,r%u,%d\n", opc2 == 0xc2 ? "c" : "g",
crfD, frA_addr, DCM);
/* setup the parameters for the long format of the two instructions */
assign( frAI64_lo, mkU64( 0 ) );
assign( gfield_mask, mkU32( DFP_G_FIELD_LONG_MASK ) );
max_exp = DFP_LONG_EXP_MAX;
min_exp = DFP_LONG_EXP_MIN;
assign( exponent, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD64,
mkexpr( frA ) ) ) ) );
assign( significand64,
unop( Iop_ReinterpI64asD64,
mkU64( 0x2234000000000001ULL ) ) ); // dfp 1.0
assign( exp_min_normal,
unop( Iop_ReinterpI64asD64, mkU64( 398 - 383 ) ) );
assign( min_subnormalD64,
binop( Iop_InsertExpD64,
mkexpr( exp_min_normal ),
mkexpr( significand64 ) ) );
assign( ccIR_subnormal,
binop( Iop_CmpD64,
mkexpr( abs_frA ),
mkexpr( min_subnormalD64 ) ) );
/* compare absolute value of frA with zero */
assign( ccIR_zero,
binop( Iop_CmpD64,
mkexpr( abs_frA ),
unop( Iop_ReinterpI64asD64,
mkU64( 0x2238000000000000ULL ) ) ) );
/* size = DFP_LONG; JRS: unused */
break;
case 0x3F: // dtstdcq, dtstdgq
DIP("dtstd%sq %u,r%u,%d\n", opc2 == 0xc2 ? "c" : "g",
crfD, frA_addr, DCM);
/* setup the parameters for the extended format of the
* two instructions
*/
assign( frAI64_lo, unop( Iop_ReinterpD64asI64,
getDReg( frA_addr+1 ) ) );
assign( gfield_mask, mkU32( DFP_G_FIELD_EXTND_MASK ) );
max_exp = DFP_EXTND_EXP_MAX;
min_exp = DFP_EXTND_EXP_MIN;
assign( exponent, unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
unop( Iop_ExtractExpD128,
getDReg_pair( frA_addr) ) ) ) );
/* create quand exponent for minimum normal number */
assign( exp_min_normal,
unop( Iop_ReinterpI64asD64, mkU64( 6176 - 6143 ) ) );
assign( significand128,
unop( Iop_D64toD128,
unop( Iop_ReinterpI64asD64,
mkU64( 0x2234000000000001ULL ) ) ) ); // dfp 1.0
assign( min_subnormalD128,
binop( Iop_InsertExpD128,
mkexpr( exp_min_normal ),
mkexpr( significand128 ) ) );
assign( ccIR_subnormal,
binop( Iop_CmpD128,
binop( Iop_D64HLtoD128,
unop( Iop_ReinterpI64asD64,
binop( Iop_And64,
unop( Iop_ReinterpD64asI64,
mkexpr( frA ) ),
mkU64( 0x7FFFFFFFFFFFFFFFULL ) ) ),
getDReg( frA_addr+1 ) ),
mkexpr( min_subnormalD128 ) ) );
assign( ccIR_zero,
binop( Iop_CmpD128,
binop( Iop_D64HLtoD128,
mkexpr( abs_frA ),
getDReg( frA_addr+1 ) ),
unop( Iop_D64toD128,
unop( Iop_ReinterpI64asD64,
mkU64( 0x0ULL ) ) ) ) );
/* size = DFP_EXTND; JRS:unused */
break;
default:
vex_printf("dis_dfp_class_test(ppc)(opc2)\n");
return False;
}
/* The G-field is in the upper 32-bits. The I64 logical operations
* do not seem to be supported in 32-bit mode so keep things as 32-bit
* operations.
*/
assign( gfield, binop( Iop_And32,
mkexpr( gfield_mask ),
unop( Iop_64HIto32,
mkexpr(frAI64_hi) ) ) );
/* There is a lot of code that is the same to do the class and group
* instructions. Later there is an if statement to handle the specific
* instruction.
*
* Will be using I32 values, compares, shifts and logical operations for
* this code as the 64-bit compare, shifts, logical operations are not
* supported in 32-bit mode.
*/
/* Check the bits for Infinity, QNaN or Signaling NaN */
assign( infinity_true,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
binop( Iop_And32,
mkU32( 0x7C000000 ),
mkexpr( gfield ) ),
mkU32( 0x78000000 ) ) ) );
assign( SNaN_true,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
binop( Iop_And32,
mkU32( 0x7E000000 ),
mkexpr( gfield ) ),
mkU32( 0x7E000000 ) ) ) );
assign( QNaN_true,
binop( Iop_And32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
binop( Iop_And32,
mkU32( 0x7E000000 ),
mkexpr( gfield ) ),
mkU32( 0x7C000000 ) ) ),
unop( Iop_Not32,
mkexpr( SNaN_true ) ) ) );
assign( zero_true,
binop( Iop_And32,
unop(Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( ccIR_zero ),
mkU32( 0x40 ) ) ), // ccIR code for Equal
unop( Iop_Not32,
binop( Iop_Or32,
mkexpr( infinity_true ),
binop( Iop_Or32,
mkexpr( QNaN_true ),
mkexpr( SNaN_true ) ) ) ) ) );
/* Do compare of frA the minimum normal value. Comparison is size
* depenent and was done above to get the ccIR value.
*/
assign( subnormal_true,
binop( Iop_And32,
binop( Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( ccIR_subnormal ),
mkU32( 0x40 ) ) ), // ccIR code for Equal
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( ccIR_subnormal ),
mkU32( 0x1 ) ) ) ), // ccIR code for LT
unop( Iop_Not32,
binop( Iop_Or32,
binop( Iop_Or32,
mkexpr( infinity_true ),
mkexpr( zero_true) ),
binop( Iop_Or32,
mkexpr( QNaN_true ),
mkexpr( SNaN_true ) ) ) ) ) );
/* Normal number is not subnormal, infinity, NaN or Zero */
assign( normal_true,
unop( Iop_Not32,
binop( Iop_Or32,
binop( Iop_Or32,
mkexpr( infinity_true ),
mkexpr( zero_true ) ),
binop( Iop_Or32,
mkexpr( subnormal_true ),
binop( Iop_Or32,
mkexpr( QNaN_true ),
mkexpr( SNaN_true ) ) ) ) ) );
/* Calculate the DCM bit field based on the tests for the specific
* instruction
*/
if (opc2 == 0xC2) { // dtstdc, dtstdcq
/* DCM[0:5] Bit Data Class definition
* 0 Zero
* 1 Subnormal
* 2 Normal
* 3 Infinity
* 4 Quiet NaN
* 5 Signaling NaN
*/
assign( dcm0, binop( Iop_Shl32,
mkexpr( zero_true ),
mkU8( 5 ) ) );
assign( dcm1, binop( Iop_Shl32,
binop( Iop_And32,
mkexpr( subnormal_true ),
mkU32( 1 ) ),
mkU8( 4 ) ) );
assign( dcm2, binop( Iop_Shl32,
binop( Iop_And32,
mkexpr( normal_true ),
mkU32( 1 ) ),
mkU8( 3 ) ) );
assign( dcm3, binop( Iop_Shl32,
binop( Iop_And32,
mkexpr( infinity_true),
mkU32( 1 ) ),
mkU8( 2 ) ) );
assign( dcm4, binop( Iop_Shl32,
binop( Iop_And32,
mkexpr( QNaN_true ),
mkU32( 1 ) ),
mkU8( 1 ) ) );
assign( dcm5, binop( Iop_And32, mkexpr( SNaN_true), mkU32( 1 ) ) );
} else if (opc2 == 0xE2) { // dtstdg, dtstdgq
/* check if the exponent is extreme */
assign( extreme_true, binop( Iop_Or32,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( exponent ),
mkU32( max_exp ) ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( exponent ),
mkU32( min_exp ) ) ) ) );
/* Check if LMD is zero */
Get_lmd( &lmd, binop( Iop_Shr32,
mkexpr( gfield ), mkU8( 31 - 5 ) ) );
assign( lmd_zero_true, unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( lmd ),
mkU32( 0 ) ) ) );
/* DCM[0:5] Bit Data Class definition
* 0 Zero with non-extreme exponent
* 1 Zero with extreme exponent
* 2 Subnormal or (Normal with extreme exponent)
* 3 Normal with non-extreme exponent and
* leftmost zero digit in significand
* 4 Normal with non-extreme exponent and
* leftmost nonzero digit in significand
* 5 Special symbol (Infinity, QNaN, or SNaN)
*/
assign( dcm0, binop( Iop_Shl32,
binop( Iop_And32,
binop( Iop_And32,
unop( Iop_Not32,
mkexpr( extreme_true ) ),
mkexpr( zero_true ) ),
mkU32( 0x1 ) ),
mkU8( 5 ) ) );
assign( dcm1, binop( Iop_Shl32,
binop( Iop_And32,
binop( Iop_And32,
mkexpr( extreme_true ),
mkexpr( zero_true ) ),
mkU32( 0x1 ) ),
mkU8( 4 ) ) );
assign( dcm2, binop( Iop_Shl32,
binop( Iop_And32,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( extreme_true ),
mkexpr( normal_true ) ),
mkexpr( subnormal_true ) ),
mkU32( 0x1 ) ),
mkU8( 3 ) ) );
assign( dcm3, binop( Iop_Shl32,
binop( Iop_And32,
binop( Iop_And32,
binop( Iop_And32,
unop( Iop_Not32,
mkexpr( extreme_true ) ),
mkexpr( normal_true ) ),
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
mkexpr( lmd ),
mkU32( 0 ) ) ) ),
mkU32( 0x1 ) ),
mkU8( 2 ) ) );
assign( dcm4, binop( Iop_Shl32,
binop( Iop_And32,
binop( Iop_And32,
binop( Iop_And32,
unop( Iop_Not32,
mkexpr( extreme_true ) ),
mkexpr( normal_true ) ),
unop( Iop_1Sto32,
binop( Iop_CmpNE32,
mkexpr( lmd ),
mkU32( 0 ) ) ) ),
mkU32( 0x1 ) ),
mkU8( 1 ) ) );
assign( dcm5, binop( Iop_And32,
binop( Iop_Or32,
mkexpr( SNaN_true),
binop( Iop_Or32,
mkexpr( QNaN_true),
mkexpr( infinity_true) ) ),
mkU32( 0x1 ) ) );
}
/* create DCM field */
assign( DCM_calc,
binop( Iop_Or32,
mkexpr( dcm0 ),
binop( Iop_Or32,
mkexpr( dcm1 ),
binop( Iop_Or32,
mkexpr( dcm2 ),
binop( Iop_Or32,
mkexpr( dcm3 ),
binop( Iop_Or32,
mkexpr( dcm4 ),
mkexpr( dcm5 ) ) ) ) ) ) );
/* Get the sign of the DFP number, ignore sign for QNaN */
assign( sign,
unop( Iop_1Uto32,
binop( Iop_CmpEQ32,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frAI64_hi ) ),
mkU8( 63 - 32 ) ),
mkU32( 1 ) ) ) );
/* This instruction generates a four bit field to be stored in the
* condition code register. The condition code register consists of 7
* fields. The field to be written to is specified by the BF (AKA crfD)
* field.
*
* The field layout is as follows:
*
* Field Meaning
* 0000 Operand positive with no match
* 0100 Operand positive with at least one match
* 0001 Operand negative with no match
* 0101 Operand negative with at least one match
*/
assign( field, binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( sign ),
mkU8( 3 ) ),
binop( Iop_Shl32,
unop( Iop_1Uto32,
binop( Iop_CmpNE32,
binop( Iop_And32,
mkU32( DCM ),
mkexpr( DCM_calc ) ),
mkU32( 0 ) ) ),
mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR, mkexpr( field ), crfD );
return True;
}
static Bool dis_dfp_bcd(UInt theInstr) {
UInt opc2 = ifieldOPClo10( theInstr );
ULong sp = IFIELD(theInstr, 19, 2);
ULong s = IFIELD(theInstr, 20, 1);
UChar frT_addr = ifieldRegDS( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
IRTemp frB = newTemp( Ity_D64 );
IRTemp frBI64 = newTemp( Ity_I64 );
IRTemp result = newTemp( Ity_I64 );
IRTemp resultD64 = newTemp( Ity_D64 );
IRTemp bcd64 = newTemp( Ity_I64 );
IRTemp bcd_u = newTemp( Ity_I32 );
IRTemp bcd_l = newTemp( Ity_I32 );
IRTemp dbcd_u = newTemp( Ity_I32 );
IRTemp dbcd_l = newTemp( Ity_I32 );
IRTemp lmd = newTemp( Ity_I32 );
assign( frB, getDReg( frB_addr ) );
assign( frBI64, unop( Iop_ReinterpD64asI64, mkexpr( frB ) ) );
switch ( opc2 ) {
case 0x142: // ddedpd DFP Decode DPD to BCD
DIP( "ddedpd %llu,r%u,r%u\n", sp, frT_addr, frB_addr );
assign( bcd64, unop( Iop_DPBtoBCD, mkexpr( frBI64 ) ) );
assign( bcd_u, unop( Iop_64HIto32, mkexpr( bcd64 ) ) );
assign( bcd_l, unop( Iop_64to32, mkexpr( bcd64 ) ) );
if ( ( sp == 0 ) || ( sp == 1 ) ) {
/* Unsigned BCD string */
Get_lmd( &lmd,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64 ) ),
mkU8( 31 - 5 ) ) ); // G-field[0:4]
assign( result,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32, mkexpr( lmd ), mkU8( 28 ) ),
mkexpr( bcd_u ) ),
mkexpr( bcd_l ) ) );
} else {
/* Signed BCD string, the cases for sp 2 and 3 only differ in how
* the positive and negative values are encoded in the least
* significant bits.
*/
IRTemp sign = newTemp( Ity_I32 );
if (sp == 2) {
/* Positive sign = 0xC, negative sign = 0xD */
assign( sign,
binop( Iop_Or32,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64 ) ),
mkU8( 31 ) ),
mkU32( 0xC ) ) );
} else if ( sp == 3 ) {
/* Positive sign = 0xF, negative sign = 0xD */
IRTemp tmp32 = newTemp( Ity_I32 );
/* Complement sign bit then OR into bit position 1 */
assign( tmp32,
binop( Iop_Xor32,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64 ) ),
mkU8( 30 ) ),
mkU32( 0x2 ) ) );
assign( sign, binop( Iop_Or32, mkexpr( tmp32 ), mkU32( 0xD ) ) );
} else {
vpanic( "The impossible happened: dis_dfp_bcd(ppc), undefined SP field" );
}
/* Put sign in bottom 4 bits, move most significant 4-bits from
* bcd_l to bcd_u.
*/
assign( result,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shr32,
mkexpr( bcd_l ),
mkU8( 28 ) ),
binop( Iop_Shl32,
mkexpr( bcd_u ),
mkU8( 4 ) ) ),
binop( Iop_Or32,
mkexpr( sign ),
binop( Iop_Shl32,
mkexpr( bcd_l ),
mkU8( 4 ) ) ) ) );
}
putDReg( frT_addr, unop( Iop_ReinterpI64asD64, mkexpr( result ) ) );
break;
case 0x342: // denbcd DFP Encode BCD to DPD
{
IRTemp valid_mask = newTemp( Ity_I32 );
IRTemp invalid_mask = newTemp( Ity_I32 );
IRTemp without_lmd = newTemp( Ity_I64 );
IRTemp tmp64 = newTemp( Ity_I64 );
IRTemp dbcd64 = newTemp( Ity_I64 );
IRTemp left_exp = newTemp( Ity_I32 );
IRTemp g0_4 = newTemp( Ity_I32 );
DIP( "denbcd %llu,r%u,r%u\n", s, frT_addr, frB_addr );
if ( s == 0 ) {
/* Unsigned BCD string */
assign( dbcd64, unop( Iop_BCDtoDPB, mkexpr(frBI64 ) ) );
assign( dbcd_u, unop( Iop_64HIto32, mkexpr( dbcd64 ) ) );
assign( dbcd_l, unop( Iop_64to32, mkexpr( dbcd64 ) ) );
assign( lmd,
binop( Iop_Shr32,
binop( Iop_And32,
unop( Iop_64HIto32, mkexpr( frBI64 ) ),
mkU32( 0xF0000000 ) ),
mkU8( 28 ) ) );
assign( invalid_mask,
bcd_digit_inval( unop( Iop_64HIto32, mkexpr( frBI64 ) ),
unop( Iop_64to32, mkexpr( frBI64 ) ) ) );
assign( valid_mask, unop( Iop_Not32, mkexpr( invalid_mask ) ) );
assign( without_lmd,
unop( Iop_ReinterpD64asI64,
binop( Iop_InsertExpD64,
unop( Iop_ReinterpI64asD64,
mkU64( DFP_LONG_BIAS ) ),
unop( Iop_ReinterpI64asD64,
binop( Iop_32HLto64,
mkexpr( dbcd_u ),
mkexpr( dbcd_l ) ) ) ) ) );
assign( left_exp,
binop( Iop_Shr32,
binop( Iop_And32,
unop( Iop_64HIto32, mkexpr( without_lmd ) ),
mkU32( 0x60000000 ) ),
mkU8( 29 ) ) );
assign( g0_4,
binop( Iop_Shl32,
Gfield_encoding( mkexpr( left_exp ), mkexpr( lmd ) ),
mkU8( 26 ) ) );
assign( tmp64,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
unop( Iop_64HIto32,
mkexpr( without_lmd ) ),
mkU32( 0x83FFFFFF ) ),
mkexpr( g0_4 ) ),
unop( Iop_64to32, mkexpr( without_lmd ) ) ) );
} else if ( s == 1 ) {
IRTemp sign = newTemp( Ity_I32 );
IRTemp sign_bit = newTemp( Ity_I32 );
IRTemp pos_sign_mask = newTemp( Ity_I32 );
IRTemp neg_sign_mask = newTemp( Ity_I32 );
IRTemp tmp = newTemp( Ity_I64 );
/* Signed BCD string, least significant 4 bits are sign bits
* positive sign = 0xC, negative sign = 0xD
*/
assign( tmp, unop( Iop_BCDtoDPB,
binop( Iop_32HLto64,
binop( Iop_Shr32,
unop( Iop_64HIto32,
mkexpr( frBI64 ) ),
mkU8( 4 ) ),
binop( Iop_Or32,
binop( Iop_Shr32,
unop( Iop_64to32,
mkexpr( frBI64 ) ),
mkU8( 4 ) ),
binop( Iop_Shl32,
unop( Iop_64HIto32,
mkexpr( frBI64 ) ),
mkU8( 28 ) ) ) ) ) );
assign( dbcd_u, unop( Iop_64HIto32, mkexpr( tmp ) ) );
assign( dbcd_l, unop( Iop_64to32, mkexpr( tmp ) ) );
/* Get the sign of the BCD string. */
assign( sign,
binop( Iop_And32,
unop( Iop_64to32, mkexpr( frBI64 ) ),
mkU32( 0xF ) ) );
assign( neg_sign_mask, Generate_neg_sign_mask( mkexpr( sign ) ) );
assign( pos_sign_mask, Generate_pos_sign_mask( mkexpr( sign ) ) );
assign( sign_bit,
Generate_sign_bit( mkexpr( pos_sign_mask ),
mkexpr( neg_sign_mask ) ) );
/* Check for invalid sign and BCD digit. Don't check the bottom
* four bits of bcd_l as that is the sign value.
*/
assign( invalid_mask,
Generate_inv_mask(
bcd_digit_inval( unop( Iop_64HIto32,
mkexpr( frBI64 ) ),
binop( Iop_Shr32,
unop( Iop_64to32,
mkexpr( frBI64 ) ),
mkU8( 4 ) ) ),
mkexpr( pos_sign_mask ),
mkexpr( neg_sign_mask ) ) );
assign( valid_mask, unop( Iop_Not32, mkexpr( invalid_mask ) ) );
/* Generate the result assuming the sign value was valid. */
assign( tmp64,
unop( Iop_ReinterpD64asI64,
binop( Iop_InsertExpD64,
unop( Iop_ReinterpI64asD64,
mkU64( DFP_LONG_BIAS ) ),
unop( Iop_ReinterpI64asD64,
binop( Iop_32HLto64,
binop( Iop_Or32,
mkexpr( dbcd_u ),
mkexpr( sign_bit ) ),
mkexpr( dbcd_l ) ) ) ) ) );
}
/* Generate the value to store depending on the validity of the
* sign value and the validity of the BCD digits.
*/
assign( resultD64,
unop( Iop_ReinterpI64asD64,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64HIto32,
mkexpr( tmp64 ) ) ),
binop( Iop_And32,
mkU32( 0x7C000000 ),
mkexpr( invalid_mask ) ) ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64to32, mkexpr( tmp64 ) ) ),
binop( Iop_And32,
mkU32( 0x0 ),
mkexpr( invalid_mask ) ) ) ) ) );
putDReg( frT_addr, mkexpr( resultD64 ) );
}
break;
default:
vpanic( "ERROR: dis_dfp_bcd(ppc), undefined opc2 case " );
return False;
}
return True;
}
static Bool dis_dfp_bcdq( UInt theInstr )
{
UInt opc2 = ifieldOPClo10( theInstr );
ULong sp = IFIELD(theInstr, 19, 2);
ULong s = IFIELD(theInstr, 20, 1);
IRTemp frB_hi = newTemp( Ity_D64 );
IRTemp frB_lo = newTemp( Ity_D64 );
IRTemp frBI64_hi = newTemp( Ity_I64 );
IRTemp frBI64_lo = newTemp( Ity_I64 );
UChar frT_addr = ifieldRegDS( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
IRTemp lmd = newTemp( Ity_I32 );
IRTemp result_hi = newTemp( Ity_I64 );
IRTemp result_lo = newTemp( Ity_I64 );
assign( frB_hi, getDReg( frB_addr ) );
assign( frB_lo, getDReg( frB_addr + 1 ) );
assign( frBI64_hi, unop( Iop_ReinterpD64asI64, mkexpr( frB_hi ) ) );
assign( frBI64_lo, unop( Iop_ReinterpD64asI64, mkexpr( frB_lo ) ) );
switch ( opc2 ) {
case 0x142: // ddedpdq DFP Decode DPD to BCD
{
IRTemp low_60_u = newTemp( Ity_I32 );
IRTemp low_60_l = newTemp( Ity_I32 );
IRTemp mid_60_u = newTemp( Ity_I32 );
IRTemp mid_60_l = newTemp( Ity_I32 );
IRTemp top_12_l = newTemp( Ity_I32 );
DIP( "ddedpdq %llu,r%u,r%u\n", sp, frT_addr, frB_addr );
/* Note, instruction only stores the lower 32 BCD digits in
* the result
*/
Generate_132_bit_bcd_string( mkexpr( frBI64_hi ),
mkexpr( frBI64_lo ),
&top_12_l,
&mid_60_u,
&mid_60_l,
&low_60_u,
&low_60_l );
if ( ( sp == 0 ) || ( sp == 1 ) ) {
/* Unsigned BCD string */
assign( result_hi,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( top_12_l ),
mkU8( 24 ) ),
binop( Iop_Shr32,
mkexpr( mid_60_u ),
mkU8( 4 ) ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( mid_60_u ),
mkU8( 28 ) ),
binop( Iop_Shr32,
mkexpr( mid_60_l ),
mkU8( 4 ) ) ) ) );
assign( result_lo,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( mid_60_l ),
mkU8( 28 ) ),
mkexpr( low_60_u ) ),
mkexpr( low_60_l ) ) );
} else {
/* Signed BCD string, the cases for sp 2 and 3 only differ in how
* the positive and negative values are encoded in the least
* significant bits.
*/
IRTemp sign = newTemp( Ity_I32 );
if ( sp == 2 ) {
/* Positive sign = 0xC, negative sign = 0xD */
assign( sign,
binop( Iop_Or32,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64_hi ) ),
mkU8( 31 ) ),
mkU32( 0xC ) ) );
} else if ( sp == 3 ) {
IRTemp tmp32 = newTemp( Ity_I32 );
/* Positive sign = 0xF, negative sign = 0xD.
* Need to complement sign bit then OR into bit position 1.
*/
assign( tmp32,
binop( Iop_Xor32,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64_hi ) ),
mkU8( 30 ) ),
mkU32( 0x2 ) ) );
assign( sign, binop( Iop_Or32, mkexpr( tmp32 ), mkU32( 0xD ) ) );
} else {
vpanic( "The impossible happened: dis_dfp_bcd(ppc), undefined SP field" );
}
assign( result_hi,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( top_12_l ),
mkU8( 28 ) ),
mkexpr( mid_60_u ) ),
mkexpr( mid_60_l ) ) );
assign( result_lo,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( low_60_u ),
mkU8( 4 ) ),
binop( Iop_Shr32,
mkexpr( low_60_l ),
mkU8( 28 ) ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( low_60_l ),
mkU8( 4 ) ),
mkexpr( sign ) ) ) );
}
putDReg( frT_addr, unop( Iop_ReinterpI64asD64, mkexpr( result_hi ) ) );
putDReg( frT_addr + 1,
unop( Iop_ReinterpI64asD64, mkexpr( result_lo ) ) );
}
break;
case 0x342: // denbcdq DFP Encode BCD to DPD
{
IRTemp valid_mask = newTemp( Ity_I32 );
IRTemp invalid_mask = newTemp( Ity_I32 );
IRTemp result128 = newTemp( Ity_D128 );
IRTemp dfp_significand = newTemp( Ity_D128 );
IRTemp tmp_hi = newTemp( Ity_I64 );
IRTemp tmp_lo = newTemp( Ity_I64 );
IRTemp dbcd_top_l = newTemp( Ity_I32 );
IRTemp dbcd_mid_u = newTemp( Ity_I32 );
IRTemp dbcd_mid_l = newTemp( Ity_I32 );
IRTemp dbcd_low_u = newTemp( Ity_I32 );
IRTemp dbcd_low_l = newTemp( Ity_I32 );
IRTemp bcd_top_8 = newTemp( Ity_I64 );
IRTemp bcd_mid_60 = newTemp( Ity_I64 );
IRTemp bcd_low_60 = newTemp( Ity_I64 );
IRTemp sign_bit = newTemp( Ity_I32 );
IRTemp tmptop10 = newTemp( Ity_I64 );
IRTemp tmpmid50 = newTemp( Ity_I64 );
IRTemp tmplow50 = newTemp( Ity_I64 );
IRTemp inval_bcd_digit_mask = newTemp( Ity_I32 );
DIP( "denbcd %llu,r%u,r%u\n", s, frT_addr, frB_addr );
if ( s == 0 ) {
/* Unsigned BCD string */
assign( sign_bit, mkU32( 0 ) ); // set to zero for unsigned string
assign( bcd_top_8,
binop( Iop_32HLto64,
mkU32( 0 ),
binop( Iop_And32,
binop( Iop_Shr32,
unop( Iop_64HIto32,
mkexpr( frBI64_hi ) ),
mkU8( 24 ) ),
mkU32( 0xFF ) ) ) );
assign( bcd_mid_60,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_Shr32,
unop( Iop_64to32,
mkexpr( frBI64_hi ) ),
mkU8( 28 ) ),
binop( Iop_Shl32,
unop( Iop_64HIto32,
mkexpr( frBI64_hi ) ),
mkU8( 4 ) ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
unop( Iop_64to32,
mkexpr( frBI64_hi ) ),
mkU8( 4 ) ),
binop( Iop_Shr32,
unop( Iop_64HIto32,
mkexpr( frBI64_lo ) ),
mkU8( 28 ) ) ) ) );
/* Note, the various helper functions ignores top 4-bits */
assign( bcd_low_60, mkexpr( frBI64_lo ) );
assign( tmptop10, unop( Iop_BCDtoDPB, mkexpr( bcd_top_8 ) ) );
assign( dbcd_top_l, unop( Iop_64to32, mkexpr( tmptop10 ) ) );
assign( tmpmid50, unop( Iop_BCDtoDPB, mkexpr( bcd_mid_60 ) ) );
assign( dbcd_mid_u, unop( Iop_64HIto32, mkexpr( tmpmid50 ) ) );
assign( dbcd_mid_l, unop( Iop_64to32, mkexpr( tmpmid50 ) ) );
assign( tmplow50, unop( Iop_BCDtoDPB, mkexpr( bcd_low_60 ) ) );
assign( dbcd_low_u, unop( Iop_64HIto32, mkexpr( tmplow50 ) ) );
assign( dbcd_low_l, unop( Iop_64to32, mkexpr( tmplow50 ) ) );
/* The entire BCD string fits in lower 110-bits. The LMD = 0,
* value is not part of the final result. Only the right most
* BCD digits are stored.
*/
assign( lmd, mkU32( 0 ) );
assign( invalid_mask,
binop( Iop_Or32,
bcd_digit_inval( mkU32( 0 ),
unop( Iop_64to32,
mkexpr( bcd_top_8 ) ) ),
binop( Iop_Or32,
bcd_digit_inval( unop( Iop_64HIto32,
mkexpr( bcd_mid_60 ) ),
unop( Iop_64to32,
mkexpr( bcd_mid_60 ) ) ),
bcd_digit_inval( unop( Iop_64HIto32,
mkexpr( bcd_low_60 ) ),
unop( Iop_64to32,
mkexpr( bcd_low_60 ) )
) ) ) );
} else if ( s == 1 ) {
IRTemp sign = newTemp( Ity_I32 );
IRTemp zero = newTemp( Ity_I32 );
IRTemp pos_sign_mask = newTemp( Ity_I32 );
IRTemp neg_sign_mask = newTemp( Ity_I32 );
/* The sign of the BCD string is stored in lower 4 bits */
assign( sign,
binop( Iop_And32,
unop( Iop_64to32, mkexpr( frBI64_lo ) ),
mkU32( 0xF ) ) );
assign( neg_sign_mask, Generate_neg_sign_mask( mkexpr( sign ) ) );
assign( pos_sign_mask, Generate_pos_sign_mask( mkexpr( sign ) ) );
assign( sign_bit,
Generate_sign_bit( mkexpr( pos_sign_mask ),
mkexpr( neg_sign_mask ) ) );
/* Generate the value assuminig the sign and BCD digits are vaild */
assign( bcd_top_8,
binop( Iop_32HLto64,
mkU32( 0x0 ),
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64_hi ) ),
mkU8( 28 ) ) ) );
/* The various helper routines ignore the upper 4-bits */
assign( bcd_mid_60, mkexpr( frBI64_hi ) );
/* Remove bottom four sign bits */
assign( bcd_low_60,
binop( Iop_32HLto64,
binop( Iop_Shr32,
unop( Iop_64HIto32,
mkexpr( frBI64_lo ) ),
mkU8( 4 ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
unop( Iop_64HIto32,
mkexpr( frBI64_lo ) ),
mkU8( 28 ) ),
binop( Iop_Shr32,
unop( Iop_64to32,
mkexpr( frBI64_lo ) ),
mkU8( 4 ) ) ) ) );
assign( tmptop10, unop( Iop_BCDtoDPB, mkexpr(bcd_top_8 ) ) );
assign( dbcd_top_l, unop( Iop_64to32, mkexpr( tmptop10 ) ) );
assign( tmpmid50, unop( Iop_BCDtoDPB, mkexpr(bcd_mid_60 ) ) );
assign( dbcd_mid_u, unop( Iop_64HIto32, mkexpr( tmpmid50 ) ) );
assign( dbcd_mid_l, unop( Iop_64to32, mkexpr( tmpmid50 ) ) );
assign( tmplow50, unop( Iop_BCDtoDPB, mkexpr( bcd_low_60 ) ) );
assign( dbcd_low_u, unop( Iop_64HIto32, mkexpr( tmplow50 ) ) );
assign( dbcd_low_l, unop( Iop_64to32, mkexpr( tmplow50 ) ) );
/* The entire BCD string fits in lower 110-bits. The LMD value
* is not stored in the final result for the DFP Long instruction.
*/
assign( lmd, mkU32( 0 ) );
/* Check for invalid sign and invalid BCD digit. Don't check the
* bottom four bits of frBI64_lo as that is the sign value.
*/
assign( zero, mkU32( 0 ) );
assign( inval_bcd_digit_mask,
binop( Iop_Or32,
bcd_digit_inval( mkexpr( zero ),
unop( Iop_64to32,
mkexpr( bcd_top_8 ) ) ),
binop( Iop_Or32,
bcd_digit_inval( unop( Iop_64HIto32,
mkexpr( bcd_mid_60 ) ),
unop( Iop_64to32,
mkexpr( bcd_mid_60 ) ) ),
bcd_digit_inval( unop( Iop_64HIto32,
mkexpr( frBI64_lo ) ),
binop( Iop_Shr32,
unop( Iop_64to32,
mkexpr( frBI64_lo ) ),
mkU8( 4 ) ) ) ) ) );
assign( invalid_mask,
Generate_inv_mask( mkexpr( inval_bcd_digit_mask ),
mkexpr( pos_sign_mask ),
mkexpr( neg_sign_mask ) ) );
}
assign( valid_mask, unop( Iop_Not32, mkexpr( invalid_mask ) ) );
/* Calculate the value of the result assuming sign and BCD digits
* are all valid.
*/
assign( dfp_significand,
binop( Iop_D64HLtoD128,
unop( Iop_ReinterpI64asD64,
binop( Iop_32HLto64,
binop( Iop_Or32,
mkexpr( sign_bit ),
mkexpr( dbcd_top_l ) ),
binop( Iop_Or32,
binop( Iop_Shl32,
mkexpr( dbcd_mid_u ),
mkU8( 18 ) ),
binop( Iop_Shr32,
mkexpr( dbcd_mid_l ),
mkU8( 14 ) ) ) ) ),
unop( Iop_ReinterpI64asD64,
binop( Iop_32HLto64,
binop( Iop_Or32,
mkexpr( dbcd_low_u ),
binop( Iop_Shl32,
mkexpr( dbcd_mid_l ),
mkU8( 18 ) ) ),
mkexpr( dbcd_low_l ) ) ) ) );
/* Break the result back down to 32-bit chunks and replace chunks.
* If there was an invalid BCD digit or invalid sign value, replace
* the calculated result with the invalid bit string.
*/
assign( result128,
binop( Iop_InsertExpD128,
unop( Iop_ReinterpI64asD64, mkU64( DFP_EXTND_BIAS ) ),
mkexpr( dfp_significand ) ) );
assign( tmp_hi,
unop( Iop_ReinterpD64asI64,
unop( Iop_D128HItoD64, mkexpr( result128 ) ) ) );
assign( tmp_lo,
unop( Iop_ReinterpD64asI64,
unop( Iop_D128LOtoD64, mkexpr( result128 ) ) ) );
assign( result_hi,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64HIto32, mkexpr( tmp_hi ) ) ),
binop( Iop_And32,
mkU32( 0x7C000000 ),
mkexpr( invalid_mask ) ) ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64to32, mkexpr( tmp_hi ) ) ),
binop( Iop_And32,
mkU32( 0x0 ),
mkexpr( invalid_mask ) ) ) ) );
assign( result_lo,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64HIto32, mkexpr( tmp_lo ) ) ),
binop( Iop_And32,
mkU32( 0x0 ),
mkexpr( invalid_mask ) ) ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( valid_mask ),
unop( Iop_64to32, mkexpr( tmp_lo ) ) ),
binop( Iop_And32,
mkU32( 0x0 ),
mkexpr( invalid_mask ) ) ) ) );
putDReg( frT_addr, unop( Iop_ReinterpI64asD64, mkexpr( result_hi ) ) );
putDReg( frT_addr + 1,
unop( Iop_ReinterpI64asD64, mkexpr( result_lo ) ) );
}
break;
default:
vpanic( "ERROR: dis_dfp_bcdq(ppc), undefined opc2 case " );
break;
}
return True;
}
static Bool dis_dfp_significant_digits( UInt theInstr )
{
UChar frA_addr = ifieldRegA( theInstr );
UChar frB_addr = ifieldRegB( theInstr );
IRTemp frA = newTemp( Ity_D64 );
UInt opc1 = ifieldOPC( theInstr );
IRTemp B_sig = newTemp( Ity_I8 );
IRTemp K = newTemp( Ity_I8 );
IRTemp lmd_B = newTemp( Ity_I32 );
IRTemp field = newTemp( Ity_I32 );
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) ); // AKA BF
IRTemp Unordered_true = newTemp( Ity_I32 );
IRTemp Eq_true_mask = newTemp( Ity_I32 );
IRTemp Lt_true_mask = newTemp( Ity_I32 );
IRTemp Gt_true_mask = newTemp( Ity_I32 );
IRTemp KisZero_true_mask = newTemp( Ity_I32 );
IRTemp KisZero_false_mask = newTemp( Ity_I32 );
/* Get the reference singificance stored in frA */
assign( frA, getDReg( frA_addr ) );
/* Convert from 64 bit to 8 bits in two steps. The Iop_64to8 is not
* supported in 32-bit mode.
*/
assign( K, unop( Iop_32to8,
binop( Iop_And32,
unop( Iop_64to32,
unop( Iop_ReinterpD64asI64,
mkexpr( frA ) ) ),
mkU32( 0x3F ) ) ) );
switch ( opc1 ) {
case 0x3b: // dtstsf DFP Test Significance
{
IRTemp frB = newTemp( Ity_D64 );
IRTemp frBI64 = newTemp( Ity_I64 );
IRTemp B_bcd_u = newTemp( Ity_I32 );
IRTemp B_bcd_l = newTemp( Ity_I32 );
IRTemp tmp64 = newTemp( Ity_I64 );
DIP( "dtstsf %u,r%u,r%u\n", crfD, frA_addr, frB_addr );
assign( frB, getDReg( frB_addr ) );
assign( frBI64, unop( Iop_ReinterpD64asI64, mkexpr( frB ) ) );
/* Get the BCD string for the value stored in a series of I32 values.
* Count the number of leading zeros. Subtract the number of leading
* zeros from 16 (maximum number of significant digits in DFP
* Long).
*/
Get_lmd( &lmd_B,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64 ) ),
mkU8( 31 - 5 ) ) ); // G-field[0:4]
assign( tmp64, unop( Iop_DPBtoBCD, mkexpr( frBI64 ) ) );
assign( B_bcd_u, unop( Iop_64HIto32, mkexpr( tmp64 ) ) );
assign( B_bcd_l, unop( Iop_64to32, mkexpr( tmp64 ) ) );
assign( B_sig,
binop( Iop_Sub8,
mkU8( DFP_LONG_MAX_SIG_DIGITS ),
Count_leading_zeros_60( mkexpr( lmd_B ),
mkexpr( B_bcd_u ),
mkexpr( B_bcd_l ) ) ) );
assign( Unordered_true, Check_unordered( mkexpr( frBI64 ) ) );
}
break;
case 0x3F: // dtstsfq DFP Test Significance
{
IRTemp frB_hi = newTemp( Ity_D64 );
IRTemp frB_lo = newTemp( Ity_D64 );
IRTemp frBI64_hi = newTemp( Ity_I64 );
IRTemp frBI64_lo = newTemp( Ity_I64 );
IRTemp B_low_60_u = newTemp( Ity_I32 );
IRTemp B_low_60_l = newTemp( Ity_I32 );
IRTemp B_mid_60_u = newTemp( Ity_I32 );
IRTemp B_mid_60_l = newTemp( Ity_I32 );
IRTemp B_top_12_l = newTemp( Ity_I32 );
DIP( "dtstsfq %u,r%u,r%u\n", crfD, frA_addr, frB_addr );
assign( frB_hi, getDReg( frB_addr ) );
assign( frB_lo, getDReg( frB_addr + 1 ) );
assign( frBI64_hi, unop( Iop_ReinterpD64asI64, mkexpr( frB_hi ) ) );
assign( frBI64_lo, unop( Iop_ReinterpD64asI64, mkexpr( frB_lo ) ) );
/* Get the BCD string for the value stored in a series of I32 values.
* Count the number of leading zeros. Subtract the number of leading
* zeros from 32 (maximum number of significant digits in DFP
* extended).
*/
Get_lmd( &lmd_B,
binop( Iop_Shr32,
unop( Iop_64HIto32, mkexpr( frBI64_hi ) ),
mkU8( 31 - 5 ) ) ); // G-field[0:4]
Generate_132_bit_bcd_string( mkexpr( frBI64_hi ),
mkexpr( frBI64_lo ),
&B_top_12_l,
&B_mid_60_u,
&B_mid_60_l,
&B_low_60_u,
&B_low_60_l );
assign( B_sig,
binop( Iop_Sub8,
mkU8( DFP_EXTND_MAX_SIG_DIGITS ),
Count_leading_zeros_128( mkexpr( lmd_B ),
mkexpr( B_top_12_l ),
mkexpr( B_mid_60_u ),
mkexpr( B_mid_60_l ),
mkexpr( B_low_60_u ),
mkexpr( B_low_60_l ) ) ) );
assign( Unordered_true, Check_unordered( mkexpr( frBI64_hi ) ) );
}
break;
}
/* Compare (16 - cnt[0]) against K and set the condition code field
* accordingly.
*
* The field layout is as follows:
*
* bit[3:0] Description
* 3 K != 0 and K < Number of significant digits if FRB
* 2 K != 0 and K > Number of significant digits if FRB OR K = 0
* 1 K != 0 and K = Number of significant digits if FRB
* 0 K ? Number of significant digits if FRB
*/
assign( Eq_true_mask,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
unop( Iop_8Uto32, mkexpr( K ) ),
unop( Iop_8Uto32, mkexpr( B_sig ) ) ) ) );
assign( Lt_true_mask,
unop( Iop_1Sto32,
binop( Iop_CmpLT32U,
unop( Iop_8Uto32, mkexpr( K ) ),
unop( Iop_8Uto32, mkexpr( B_sig ) ) ) ) );
assign( Gt_true_mask,
unop( Iop_1Sto32,
binop( Iop_CmpLT32U,
unop( Iop_8Uto32, mkexpr( B_sig ) ),
unop( Iop_8Uto32, mkexpr( K ) ) ) ) );
assign( KisZero_true_mask,
unop( Iop_1Sto32,
binop( Iop_CmpEQ32,
unop( Iop_8Uto32, mkexpr( K ) ),
mkU32( 0 ) ) ) );
assign( KisZero_false_mask,
unop( Iop_1Sto32,
binop( Iop_CmpNE32,
unop( Iop_8Uto32, mkexpr( K ) ),
mkU32( 0 ) ) ) );
assign( field,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( KisZero_false_mask ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( Lt_true_mask ),
mkU32( 0x8 ) ),
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( Gt_true_mask ),
mkU32( 0x4 ) ),
binop( Iop_And32,
mkexpr( Eq_true_mask ),
mkU32( 0x2 ) ) ) ) ),
binop( Iop_And32,
mkexpr( KisZero_true_mask ),
mkU32( 0x4 ) ) ) );
putGST_field( PPC_GST_CR,
binop( Iop_Or32,
binop( Iop_And32,
mkexpr( Unordered_true ),
mkU32( 0x1 ) ),
binop( Iop_And32,
unop( Iop_Not32, mkexpr( Unordered_true ) ),
mkexpr( field ) ) ),
crfD );
return True;
}
/*------------------------------------------------------------*/
/*--- AltiVec Instruction Translation ---*/
/*------------------------------------------------------------*/
/*
Altivec Cache Control Instructions (Data Streams)
*/
static Bool dis_av_datastream ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar flag_T = toUChar( IFIELD( theInstr, 25, 1 ) );
UChar flag_A = flag_T;
UChar b23to24 = toUChar( IFIELD( theInstr, 23, 2 ) );
UChar STRM = toUChar( IFIELD( theInstr, 21, 2 ) );
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
if (opc1 != 0x1F || b23to24 != 0 || b0 != 0) {
vex_printf("dis_av_datastream(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x156: // dst (Data Stream Touch, AV p115)
DIP("dst%s r%u,r%u,%d\n", flag_T ? "t" : "",
rA_addr, rB_addr, STRM);
break;
case 0x176: // dstst (Data Stream Touch for Store, AV p117)
DIP("dstst%s r%u,r%u,%d\n", flag_T ? "t" : "",
rA_addr, rB_addr, STRM);
break;
case 0x336: // dss (Data Stream Stop, AV p114)
if (rA_addr != 0 || rB_addr != 0) {
vex_printf("dis_av_datastream(ppc)(opc2,dst)\n");
return False;
}
if (flag_A == 0) {
DIP("dss %d\n", STRM);
} else {
DIP("dssall\n");
}
break;
default:
vex_printf("dis_av_datastream(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Processor Control Instructions
*/
static Bool dis_av_procctl ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
if (opc1 != 0x4) {
vex_printf("dis_av_procctl(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x604: // mfvscr (Move from VSCR, AV p129)
if (vA_addr != 0 || vB_addr != 0) {
vex_printf("dis_av_procctl(ppc)(opc2,dst)\n");
return False;
}
DIP("mfvscr v%d\n", vD_addr);
putVReg( vD_addr, unop(Iop_32UtoV128, getGST( PPC_GST_VSCR )) );
break;
case 0x644: { // mtvscr (Move to VSCR, AV p130)
IRTemp vB = newTemp(Ity_V128);
if (vD_addr != 0 || vA_addr != 0) {
vex_printf("dis_av_procctl(ppc)(opc2,dst)\n");
return False;
}
DIP("mtvscr v%d\n", vB_addr);
assign( vB, getVReg(vB_addr));
putGST( PPC_GST_VSCR, unop(Iop_V128to32, mkexpr(vB)) );
break;
}
default:
vex_printf("dis_av_procctl(ppc)(opc2)\n");
return False;
}
return True;
}
/*
* VSX scalar and vector convert instructions
*/
static Bool
dis_vx_conv ( UInt theInstr, UInt opc2 )
{
/* XX2-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT( theInstr );
UChar XB = ifieldRegXB( theInstr );
IRTemp xB, xB2;
IRTemp b3, b2, b1, b0;
xB = xB2 = IRTemp_INVALID;
if (opc1 != 0x3C) {
vex_printf( "dis_vx_conv(ppc)(instr)\n" );
return False;
}
/* Create and assign temps only as needed for the given instruction. */
switch (opc2) {
// scalar double-precision floating point argument
case 0x2B0: case 0x0b0: case 0x290: case 0x212: case 0x090:
xB = newTemp(Ity_F64);
assign( xB,
unop( Iop_ReinterpI64asF64,
unop( Iop_V128HIto64, getVSReg( XB ) ) ) );
break;
// vector double-precision floating point arguments
case 0x1b0: case 0x312: case 0x390: case 0x190: case 0x3B0:
xB = newTemp(Ity_F64);
xB2 = newTemp(Ity_F64);
assign( xB,
unop( Iop_ReinterpI64asF64,
unop( Iop_V128HIto64, getVSReg( XB ) ) ) );
assign( xB2,
unop( Iop_ReinterpI64asF64,
unop( Iop_V128to64, getVSReg( XB ) ) ) );
break;
// vector single precision or [un]signed integer word arguments
case 0x130: case 0x392: case 0x330: case 0x310: case 0x110:
case 0x1f0: case 0x1d0:
b3 = b2 = b1 = b0 = IRTemp_INVALID;
breakV128to4x32(getVSReg(XB), &b3, &b2, &b1, &b0);
break;
// vector [un]signed integer doubleword argument
case 0x3f0: case 0x370: case 0x3d0: case 0x350:
xB = newTemp(Ity_I64);
assign( xB, unop( Iop_V128HIto64, getVSReg( XB ) ) );
xB2 = newTemp(Ity_I64);
assign( xB2, unop( Iop_V128to64, getVSReg( XB ) ) );
break;
// scalar [un]signed integer doubleword argument
case 0x2F0: case 0x2D0:
xB = newTemp(Ity_I64);
assign( xB, unop( Iop_V128HIto64, getVSReg( XB ) ) );
break;
// scalar single precision argument
case 0x292: // xscvspdp
xB = newTemp(Ity_I32);
assign( xB,
unop( Iop_64HIto32, unop( Iop_V128HIto64, getVSReg( XB ) ) ) );
break;
/* Certain instructions have their complete implementation in the main switch statement
* that follows this one; thus we have a "do nothing" case for those instructions here.
*/
case 0x170: case 0x150:
break; // do nothing
default:
vex_printf( "dis_vx_conv(ppc)(opc2)\n" );
return False;
}
switch (opc2) {
case 0x2B0:
// xscvdpsxds (VSX Scalar truncate Double-Precision to integer and Convert
// to Signed Integer Doubleword format with Saturate)
DIP("xscvdpsxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128, binop( Iop_F64toI64S,
mkU32( Irrm_ZERO ),
mkexpr( xB ) ), mkU64( 0 ) ) );
break;
case 0x0b0: // xscvdpsxws (VSX Scalar truncate Double-Precision to integer and
// Convert to Signed Integer Word format with Saturate)
DIP("xscvdpsxws v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_32Sto64,
binop( Iop_F64toI32S,
mkU32( Irrm_ZERO ),
mkexpr( xB ) ) ),
mkU64( 0ULL ) ) );
break;
case 0x290: // xscvdpuxds (VSX Scalar truncate Double-Precision integer and Convert
// to Unsigned Integer Doubleword format with Saturate)
DIP("xscvdpuxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_F64toI64U,
mkU32( Irrm_ZERO ),
mkexpr( xB ) ),
mkU64( 0ULL ) ) );
break;
case 0x2F0:
// xscvsxddp (VSX Scalar Convert and round Signed Integer Doubleword to
// Double-Precision format)
DIP("xscvsxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
binop( Iop_I64StoF64, get_IR_roundingmode(),
mkexpr( xB ) ) ),
mkU64( 0 ) ) );
break;
case 0x2D0:
// xscvuxddp (VSX Scalar Convert and round Unsigned Integer Doubleword to
// Double-Precision format)
DIP("xscvuxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
binop( Iop_I64UtoF64, get_IR_roundingmode(),
mkexpr( xB ) ) ),
mkU64( 0 ) ) );
break;
case 0x1b0: // xvcvdpsxws (VSX Vector truncate Double-Precision to integer and Convert
// to Signed Integer Word format with Saturate)
{
IRTemp hiResult_32 = newTemp(Ity_I32);
IRTemp loResult_32 = newTemp(Ity_I32);
IRExpr* rmZero = mkU32(Irrm_ZERO);
DIP("xvcvdpsxws v%u,v%u\n", (UInt)XT, (UInt)XB);
assign(hiResult_32, binop(Iop_F64toI32S, rmZero, mkexpr(xB)));
assign(loResult_32, binop(Iop_F64toI32S, rmZero, mkexpr(xB2)));
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_32Sto64, mkexpr( hiResult_32 ) ),
unop( Iop_32Sto64, mkexpr( loResult_32 ) ) ) );
break;
}
case 0x130: case 0x110: // xvcvspsxws, xvcvspuxws
// (VSX Vector truncate Single-Precision to integer and
// Convert to [Un]signed Integer Word format with Saturate)
{
IRExpr * b0_result, * b1_result, * b2_result, * b3_result;
IRTemp tempResult = newTemp(Ity_V128);
IRTemp res0 = newTemp(Ity_I32);
IRTemp res1 = newTemp(Ity_I32);
IRTemp res2 = newTemp(Ity_I32);
IRTemp res3 = newTemp(Ity_I32);
IRTemp hi64 = newTemp(Ity_I64);
IRTemp lo64 = newTemp(Ity_I64);
Bool un_signed = (opc2 == 0x110);
IROp op = un_signed ? Iop_QFtoI32Ux4_RZ : Iop_QFtoI32Sx4_RZ;
DIP("xvcvsp%sxws v%u,v%u\n", un_signed ? "u" : "s", (UInt)XT, (UInt)XB);
/* The xvcvsp{s|u}xws instruction is similar to vct{s|u}xs, except if src is a NaN,
* then result is set to 0x80000000. */
assign(tempResult, unop(op, getVSReg(XB)));
assign( hi64, unop(Iop_V128HIto64, mkexpr(tempResult)) );
assign( lo64, unop(Iop_V128to64, mkexpr(tempResult)) );
assign( res3, unop(Iop_64HIto32, mkexpr(hi64)) );
assign( res2, unop(Iop_64to32, mkexpr(hi64)) );
assign( res1, unop(Iop_64HIto32, mkexpr(lo64)) );
assign( res0, unop(Iop_64to32, mkexpr(lo64)) );
b3_result = IRExpr_Mux0X(unop(Iop_1Uto8, is_NaN_32(b3)),
// else: result is from the Iop_QFtoI32{s|u}x4_RZ
mkexpr(res3),
// then: result is 0x{8|0}80000000
mkU32(un_signed ? 0x00000000 : 0x80000000));
b2_result = IRExpr_Mux0X(unop(Iop_1Uto8, is_NaN_32(b2)),
// else: result is from the Iop_QFtoI32{s|u}x4_RZ
mkexpr(res2),
// then: result is 0x{8|0}80000000
mkU32(un_signed ? 0x00000000 : 0x80000000));
b1_result = IRExpr_Mux0X(unop(Iop_1Uto8, is_NaN_32(b1)),
// else: result is from the Iop_QFtoI32{s|u}x4_RZ
mkexpr(res1),
// then: result is 0x{8|0}80000000
mkU32(un_signed ? 0x00000000 : 0x80000000));
b0_result = IRExpr_Mux0X(unop(Iop_1Uto8, is_NaN_32(b0)),
// else: result is from the Iop_QFtoI32{s|u}x4_RZ
mkexpr(res0),
// then: result is 0x{8|0}80000000
mkU32(un_signed ? 0x00000000 : 0x80000000));
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, b3_result, b2_result ),
binop( Iop_32HLto64, b1_result, b0_result ) ) );
break;
}
case 0x212: // xscvdpsp (VSX Scalar round Double-Precision to single-precision and
// Convert to Single-Precision format
DIP("xscvdpsp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
mkexpr( xB ) ) ) ),
mkU32( 0 ) ),
mkU64( 0ULL ) ) );
break;
case 0x090: // xscvdpuxws (VSX Scalar truncate Double-Precision to integer
// and Convert to Unsigned Integer Word format with Saturate)
DIP("xscvdpuxws v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
mkU32( 0 ),
binop( Iop_F64toI32U,
mkU32( Irrm_ZERO ),
mkexpr( xB ) ) ),
mkU64( 0ULL ) ) );
break;
case 0x292: // xscvspdp (VSX Scalar Convert Single-Precision to Double-Precision format)
DIP("xscvspdp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( xB ) ) ) ),
mkU64( 0ULL ) ) );
break;
case 0x312: // xvcvdpsp (VSX Vector round Double-Precision to single-precision
// and Convert to Single-Precision format)
DIP("xvcvdpsp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
mkexpr( xB ) ) ) ),
mkU32( 0 ) ),
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
mkexpr( xB2 ) ) ) ),
mkU32( 0 ) ) ) );
break;
case 0x390: // xvcvdpuxds (VSX Vector truncate Double-Precision to integer
// and Convert to Unsigned Integer Doubleword format
// with Saturate)
DIP("xvcvdpuxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_F64toI64U, mkU32( Irrm_ZERO ), mkexpr( xB ) ),
binop( Iop_F64toI64U, mkU32( Irrm_ZERO ), mkexpr( xB2 ) ) ) );
break;
case 0x190: // xvcvdpuxws (VSX Vector truncate Double-Precision to integer and
// Convert to Unsigned Integer Word format with Saturate)
DIP("xvcvdpuxws v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
binop( Iop_F64toI32U,
mkU32( Irrm_ZERO ),
mkexpr( xB ) ),
mkU32( 0 ) ),
binop( Iop_32HLto64,
binop( Iop_F64toI32U,
mkU32( Irrm_ZERO ),
mkexpr( xB2 ) ),
mkU32( 0 ) ) ) );
break;
case 0x392: // xvcvspdp (VSX Vector Convert Single-Precision to Double-Precision format)
DIP("xvcvspdp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b3 ) ) ) ),
unop( Iop_ReinterpF64asI64,
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b1 ) ) ) ) ) );
break;
case 0x330: // xvcvspsxds (VSX Vector truncate Single-Precision to integer and
// Convert to Signed Integer Doubleword format with Saturate)
DIP("xvcvspsxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_F64toI64S,
mkU32( Irrm_ZERO ),
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b3 ) ) ) ),
binop( Iop_F64toI64S,
mkU32( Irrm_ZERO ),
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b1 ) ) ) ) ) );
break;
case 0x310: // xvcvspuxds (VSX Vector truncate Single-Precision to integer and
// Convert to Unsigned Integer Doubleword format with Saturate)
DIP("xvcvspuxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_F64toI64U,
mkU32( Irrm_ZERO ),
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b3 ) ) ) ),
binop( Iop_F64toI64U,
mkU32( Irrm_ZERO ),
unop( Iop_F32toF64,
unop( Iop_ReinterpI32asF32, mkexpr( b1 ) ) ) ) ) );
break;
case 0x3B0: // xvcvdpsxds (VSX Vector truncate Double-Precision to integer and
// Convert to Signed Integer Doubleword format with Saturate)
DIP("xvcvdpsxds v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_F64toI64S, mkU32( Irrm_ZERO ), mkexpr( xB ) ),
binop( Iop_F64toI64S, mkU32( Irrm_ZERO ), mkexpr( xB2 ) ) ) );
break;
case 0x3f0: // xvcvsxddp (VSX Vector Convert and round Signed Integer Doubleword
// to Double-Precision format)
DIP("xvcvsxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
binop( Iop_I64StoF64,
get_IR_roundingmode(),
mkexpr( xB ) ) ),
unop( Iop_ReinterpF64asI64,
binop( Iop_I64StoF64,
get_IR_roundingmode(),
mkexpr( xB2 ) ) ) ) );
break;
case 0x3d0: // xvcvuxddp (VSX Vector Convert and round Unsigned Integer Doubleword
// to Double-Precision format)
DIP("xvcvuxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
binop( Iop_I64UtoF64,
get_IR_roundingmode(),
mkexpr( xB ) ) ),
unop( Iop_ReinterpF64asI64,
binop( Iop_I64UtoF64,
get_IR_roundingmode(),
mkexpr( xB2 ) ) ) ) );
break;
case 0x370: // xvcvsxdsp (VSX Vector Convert and round Signed Integer Doubleword
// to Single-Precision format)
DIP("xvcvsxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
binop( Iop_I64StoF64,
get_IR_roundingmode(),
mkexpr( xB ) ) ) ) ),
mkU32( 0 ) ),
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
binop( Iop_I64StoF64,
get_IR_roundingmode(),
mkexpr( xB2 ) ) ) ) ),
mkU32( 0 ) ) ) );
break;
case 0x350: // xvcvuxdsp (VSX Vector Convert and round Unsigned Integer Doubleword
// to Single-Precision format)
DIP("xvcvuxddp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
binop( Iop_I64UtoF64,
get_IR_roundingmode(),
mkexpr( xB ) ) ) ) ),
mkU32( 0 ) ),
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop( Iop_RoundF64toF32,
get_IR_roundingmode(),
binop( Iop_I64UtoF64,
get_IR_roundingmode(),
mkexpr( xB2 ) ) ) ) ),
mkU32( 0 ) ) ) );
break;
case 0x1f0: // xvcvsxwdp (VSX Vector Convert Signed Integer Word to Double-Precision format)
DIP("xvcvsxwdp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
binop( Iop_I64StoF64, get_IR_roundingmode(),
unop( Iop_32Sto64, mkexpr( b3 ) ) ) ),
unop( Iop_ReinterpF64asI64,
binop( Iop_I64StoF64, get_IR_roundingmode(),
unop( Iop_32Sto64, mkexpr( b1 ) ) ) ) ) );
break;
case 0x1d0: // xvcvuxwdp (VSX Vector Convert Unsigned Integer Word to Double-Precision format)
DIP("xvcvuxwdp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
binop( Iop_I64UtoF64, get_IR_roundingmode(),
unop( Iop_32Uto64, mkexpr( b3 ) ) ) ),
unop( Iop_ReinterpF64asI64,
binop( Iop_I64UtoF64, get_IR_roundingmode(),
unop( Iop_32Uto64, mkexpr( b1 ) ) ) ) ) );
break;
case 0x170: // xvcvsxwsp (VSX Vector Convert Signed Integer Word to Single-Precision format)
DIP("xvcvsxwsp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT, unop( Iop_I32StoFx4, getVSReg( XB ) ) );
break;
case 0x150: // xvcvuxwsp (VSX Vector Convert Unsigned Integer Word to Single-Precision format)
DIP("xvcvuxwsp v%u,v%u\n", (UInt)XT, (UInt)XB);
putVSReg( XT, unop( Iop_I32UtoFx4, getVSReg( XB ) ) );
break;
default:
vex_printf( "dis_vx_conv(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX vector Double Precision Floating Point Arithmetic Instructions
*/
static Bool
dis_vxv_dp_arith ( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT( theInstr );
UChar XA = ifieldRegXA( theInstr );
UChar XB = ifieldRegXB( theInstr );
IRExpr* rm = get_IR_roundingmode();
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
IRTemp frA2 = newTemp(Ity_F64);
IRTemp frB2 = newTemp(Ity_F64);
if (opc1 != 0x3C) {
vex_printf( "dis_vxv_dp_arith(ppc)(instr)\n" );
return False;
}
assign(frA, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XA ))));
assign(frB, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XB ))));
assign(frA2, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, getVSReg( XA ))));
assign(frB2, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, getVSReg( XB ))));
switch (opc2) {
case 0x1E0: // xvdivdp (VSX Vector Divide Double-Precision)
case 0x1C0: // xvmuldp (VSX Vector Multiply Double-Precision)
case 0x180: // xvadddp (VSX Vector Add Double-Precision)
case 0x1A0: // xvsubdp (VSX Vector Subtract Double-Precision)
{
IROp mOp;
Char * oper_name;
switch (opc2) {
case 0x1E0:
mOp = Iop_DivF64;
oper_name = "div";
break;
case 0x1C0:
mOp = Iop_MulF64;
oper_name = "mul";
break;
case 0x180:
mOp = Iop_AddF64;
oper_name = "add";
break;
case 0x1A0:
mOp = Iop_SubF64;
oper_name = "sub";
break;
default:
vpanic("The impossible happened: dis_vxv_dp_arith(ppc)");
}
IRTemp hiResult = newTemp(Ity_I64);
IRTemp loResult = newTemp(Ity_I64);
DIP("xv%sdp v%d,v%d,v%d\n", oper_name, (UInt)XT, (UInt)XA, (UInt)XB);
assign( hiResult,
unop( Iop_ReinterpF64asI64,
triop( mOp, rm, mkexpr( frA ), mkexpr( frB ) ) ) );
assign( loResult,
unop( Iop_ReinterpF64asI64,
triop( mOp, rm, mkexpr( frA2 ), mkexpr( frB2 ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128, mkexpr( hiResult ), mkexpr( loResult ) ) );
break;
}
case 0x196: // xvsqrtdp
{
IRTemp hiResult = newTemp(Ity_I64);
IRTemp loResult = newTemp(Ity_I64);
DIP("xvsqrtdp v%d,v%d\n", (UInt)XT, (UInt)XB);
assign( hiResult,
unop( Iop_ReinterpF64asI64,
binop( Iop_SqrtF64, rm, mkexpr( frB ) ) ) );
assign( loResult,
unop( Iop_ReinterpF64asI64,
binop( Iop_SqrtF64, rm, mkexpr( frB2 ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128, mkexpr( hiResult ), mkexpr( loResult ) ) );
break;
}
case 0x184: case 0x1A4: // xvmaddadp, xvmaddmdp (VSX Vector Multiply-Add Double-Precision)
case 0x1C4: case 0x1E4: // xvmsubadp, xvmsubmdp (VSX Vector Multiply-Subtract Double-Precision)
case 0x384: case 0x3A4: // xvnmaddadp, xvnmaddmdp (VSX Vector Negate Multiply-Add Double-Precision)
case 0x3C4: case 0x3E4: // xvnmsubadp, xvnmsubmdp (VSX Vector Negate Multiply-Subtract Double-Precision)
{
/* xvm{add|sub}mdp XT,XA,XB is element-wise equivalent to fm{add|sub} FRT,FRA,FRC,FRB with . . .
* XT == FRC
* XA == FRA
* XB == FRB
*
* and for xvm{add|sub}adp . . .
* XT == FRB
* XA == FRA
* XB == FRC
*/
Bool negate;
IROp mOp = Iop_INVALID;
Char * oper_name = NULL;
Bool mdp = False;
switch (opc2) {
case 0x184: case 0x1A4:
case 0x384: case 0x3A4:
mOp = Iop_MAddF64;
oper_name = "add";
mdp = (opc2 & 0x0FF) == 0x0A4;
break;
case 0x1C4: case 0x1E4:
case 0x3C4: case 0x3E4:
mOp = Iop_MSubF64;
oper_name = "sub";
mdp = (opc2 & 0x0FF) == 0x0E4;
break;
default:
vpanic("The impossible happened: dis_vxv_sp_arith(ppc)");
}
switch (opc2) {
case 0x384: case 0x3A4:
case 0x3C4: case 0x3E4:
negate = True;
break;
default:
negate = False;
}
IRTemp hiResult = newTemp(Ity_I64);
IRTemp loResult = newTemp(Ity_I64);
IRTemp frT = newTemp(Ity_F64);
IRTemp frT2 = newTemp(Ity_F64);
DIP("xv%sm%s%s v%d,v%d,v%d\n", negate ? "n" : "", oper_name, mdp ? "mdp" : "adp",
(UInt)XT, (UInt)XA, (UInt)XB);
assign(frT, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XT ) ) ) );
assign(frT2, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, getVSReg( XT ) ) ) );
assign( hiResult,
unop( Iop_ReinterpF64asI64,
qop( mOp,
rm,
mkexpr( frA ),
mkexpr( mdp ? frT : frB ),
mkexpr( mdp ? frB : frT ) ) ) );
assign( loResult,
unop( Iop_ReinterpF64asI64,
qop( mOp,
rm,
mkexpr( frA2 ),
mkexpr( mdp ? frT2 : frB2 ),
mkexpr( mdp ? frB2 : frT2 ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
mkexpr( negate ? getNegatedResult( hiResult )
: hiResult ),
mkexpr( negate ? getNegatedResult( loResult )
: loResult ) ) );
break;
}
case 0x1D4: // xvtsqrtdp (VSX Vector Test for software Square Root Double-Precision)
{
IRTemp frBHi_I64 = newTemp(Ity_I64);
IRTemp frBLo_I64 = newTemp(Ity_I64);
IRTemp flagsHi = newTemp(Ity_I32);
IRTemp flagsLo = newTemp(Ity_I32);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp fe_flagHi, fg_flagHi, fe_flagLo, fg_flagLo;
fe_flagHi = fg_flagHi = fe_flagLo = fg_flagLo = IRTemp_INVALID;
DIP("xvtsqrtdp cr%d,v%d\n", (UInt)crfD, (UInt)XB);
assign( frBHi_I64, unop(Iop_V128HIto64, getVSReg( XB )) );
assign( frBLo_I64, unop(Iop_V128to64, getVSReg( XB )) );
do_fp_tsqrt(frBHi_I64, False /*not single precision*/, &fe_flagHi, &fg_flagHi);
do_fp_tsqrt(frBLo_I64, False /*not single precision*/, &fe_flagLo, &fg_flagLo);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flagsHi,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flagHi), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flagHi), mkU8( 1 ) ) ) );
assign( flagsLo,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flagLo), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flagLo), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR,
binop( Iop_Or32, mkexpr( flagsHi ), mkexpr( flagsLo ) ),
crfD );
break;
}
case 0x1F4: // xvtdivdp (VSX Vector Test for software Divide Double-Precision)
{
IRTemp frBHi_I64 = newTemp(Ity_I64);
IRTemp frBLo_I64 = newTemp(Ity_I64);
IRTemp frAHi_I64 = newTemp(Ity_I64);
IRTemp frALo_I64 = newTemp(Ity_I64);
IRTemp flagsHi = newTemp(Ity_I32);
IRTemp flagsLo = newTemp(Ity_I32);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp fe_flagHi, fg_flagHi, fe_flagLo, fg_flagLo;
fe_flagHi = fg_flagHi = fe_flagLo = fg_flagLo = IRTemp_INVALID;
DIP("xvtdivdp cr%d,v%d,v%d\n", (UInt)crfD, (UInt)XA, (UInt)XB);
assign( frAHi_I64, unop(Iop_V128HIto64, getVSReg( XA )) );
assign( frALo_I64, unop(Iop_V128to64, getVSReg( XA )) );
assign( frBHi_I64, unop(Iop_V128HIto64, getVSReg( XB )) );
assign( frBLo_I64, unop(Iop_V128to64, getVSReg( XB )) );
_do_fp_tdiv(frAHi_I64, frBHi_I64, False/*dp*/, &fe_flagHi, &fg_flagHi);
_do_fp_tdiv(frALo_I64, frBLo_I64, False/*dp*/, &fe_flagLo, &fg_flagLo);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flagsHi,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flagHi), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flagHi), mkU8( 1 ) ) ) );
assign( flagsLo,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flagLo), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flagLo), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR,
binop( Iop_Or32, mkexpr( flagsHi ), mkexpr( flagsLo ) ),
crfD );
break;
}
default:
vex_printf( "dis_vxv_dp_arith(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX vector Single Precision Floating Point Arithmetic Instructions
*/
static Bool
dis_vxv_sp_arith ( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT( theInstr );
UChar XA = ifieldRegXA( theInstr );
UChar XB = ifieldRegXB( theInstr );
IRExpr* rm = get_IR_roundingmode();
IRTemp a3, a2, a1, a0;
IRTemp b3, b2, b1, b0;
IRTemp res0 = newTemp(Ity_I32);
IRTemp res1 = newTemp(Ity_I32);
IRTemp res2 = newTemp(Ity_I32);
IRTemp res3 = newTemp(Ity_I32);
a3 = a2 = a1 = a0 = IRTemp_INVALID;
b3 = b2 = b1 = b0 = IRTemp_INVALID;
if (opc1 != 0x3C) {
vex_printf( "dis_vxv_sp_arith(ppc)(instr)\n" );
return False;
}
switch (opc2) {
case 0x100: // xvaddsp (VSX Vector Add Single-Precision)
DIP("xvaddsp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop(Iop_Add32Fx4, getVSReg( XA ), getVSReg( XB )) );
break;
case 0x140: // xvmulsp (VSX Vector Multiply Single-Precision)
DIP("xvmulsp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop(Iop_Mul32Fx4, getVSReg( XA ), getVSReg( XB )) );
break;
case 0x120: // xvsubsp (VSX Vector Subtract Single-Precision)
DIP("xvsubsp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop(Iop_Sub32Fx4, getVSReg( XA ), getVSReg( XB )) );
break;
case 0x160: // xvdivsp (VSX Vector Divide Single-Precision)
{
/* Iop_Div32Fx4 is not implemented for ppc64 (in host_ppc_{isel|defs}.c.
* So there are two choices:
* 1. Implement the xvdivsp with a native insn; or
* 2. Extract the 4 single precision floats from each vector
* register inputs and perform fdivs on each pair
* I will do the latter, due to the general philosophy of
* reusing existing implementations when practical.
*/
DIP("xvdivsp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
breakV128to4xF64( getVSReg( XA ), &a3, &a2, &a1, &a0 );
breakV128to4xF64( getVSReg( XB ), &b3, &b2, &b1, &b0 );
assign( res0,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32, rm, mkexpr( a0 ), mkexpr( b0 ) ) ) ) );
assign( res1,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32, rm, mkexpr( a1 ), mkexpr( b1 ) ) ) ) );
assign( res2,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32, rm, mkexpr( a2 ), mkexpr( b2 ) ) ) ) );
assign( res3,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32, rm, mkexpr( a3 ), mkexpr( b3 ) ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkexpr( res3 ), mkexpr( res2 ) ),
binop( Iop_32HLto64, mkexpr( res1 ), mkexpr( res0 ) ) ) );
break;
}
case 0x116: // xvsqrtsp (VSX Vector Square Root Single-Precision)
{
DIP("xvsqrtsp v%d,v%d\n", (UInt)XT, (UInt)XB);
breakV128to4xF64( getVSReg( XB ), &b3, &b2, &b1, &b0 );
/* Note: The native xvsqrtsp insruction does not always give the same precision
* as what we get with Iop_SqrtF64. But it doesn't seem worthwhile to implement
* an Iop_SqrtF32 that would give us a lower precision result, albeit more true
* to the actual instruction.
*/
assign( res0,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop(Iop_SqrtF64, rm, mkexpr( b0 ) ) ) ) );
assign( res1,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop(Iop_SqrtF64, rm, mkexpr( b1 ) ) ) ) );
assign( res2,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop(Iop_SqrtF64, rm, mkexpr( b2) ) ) ) );
assign( res3,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
binop(Iop_SqrtF64, rm, mkexpr( b3 ) ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkexpr( res3 ), mkexpr( res2 ) ),
binop( Iop_32HLto64, mkexpr( res1 ), mkexpr( res0 ) ) ) );
break;
}
case 0x104: case 0x124: // xvmaddasp, xvmaddmsp (VSX Vector Multiply-Add Single-Precision)
case 0x144: case 0x164: // xvmsubasp, xvmsubmsp (VSX Vector Multiply-Subtract Single-Precision)
case 0x304: case 0x324: // xvnmaddasp, xvnmaddmsp (VSX Vector Negate Multiply-Add Single-Precision)
case 0x344: case 0x364: // xvnmsubasp, xvnmsubmsp (VSX Vector Negate Multiply-Subtract Single-Precision)
{
IRTemp t3, t2, t1, t0;
Bool msp = False;
Bool negate;
Char * oper_name = NULL;
IROp mOp = Iop_INVALID;
switch (opc2) {
case 0x104: case 0x124:
case 0x304: case 0x324:
msp = (opc2 & 0x0FF) == 0x024;
mOp = Iop_MAddF64r32;
oper_name = "madd";
break;
case 0x144: case 0x164:
case 0x344: case 0x364:
msp = (opc2 & 0x0FF) == 0x064;
mOp = Iop_MSubF64r32;
oper_name = "sub";
break;
default:
vpanic("The impossible happened: dis_vxv_sp_arith(ppc)");
}
switch (opc2) {
case 0x304: case 0x324:
case 0x344: case 0x364:
negate = True;
break;
default:
negate = False;
}
DIP("xv%sm%s%s v%d,v%d,v%d\n", negate ? "n" : "", oper_name, msp ? "msp" : "asp",
(UInt)XT, (UInt)XA, (UInt)XB);
t3 = t2 = t1 = t0 = IRTemp_INVALID;
breakV128to4xF64( getVSReg( XA ), &a3, &a2, &a1, &a0 );
breakV128to4xF64( getVSReg( XB ), &b3, &b2, &b1, &b0 );
breakV128to4xF64( getVSReg( XT ), &t3, &t2, &t1, &t0 );
assign( res0,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
qop( mOp,
rm,
mkexpr( a0 ),
mkexpr( msp ? t0 : b0 ),
mkexpr( msp ? b0 : t0 ) ) ) ) );
assign( res1,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
qop( mOp,
rm,
mkexpr( a1 ),
mkexpr( msp ? t1 : b1 ),
mkexpr( msp ? b1 : t1 ) ) ) ) );
assign( res2,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
qop( mOp,
rm,
mkexpr( a2 ),
mkexpr( msp ? t2 : b2 ),
mkexpr( msp ? b2 : t2 ) ) ) ) );
assign( res3,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
qop( mOp,
rm,
mkexpr( a3 ),
mkexpr( msp ? t3 : b3 ),
mkexpr( msp ? b3 : t3 ) ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkexpr( negate ? getNegatedResult_32( res3 ) : res3 ),
mkexpr( negate ? getNegatedResult_32( res2 ) : res2 ) ),
binop( Iop_32HLto64, mkexpr( negate ? getNegatedResult_32( res1 ) : res1 ),
mkexpr( negate ? getNegatedResult_32( res0 ) : res0 ) ) ) );
break;
}
case 0x154: // xvtsqrtsp (VSX Vector Test for software Square Root Single-Precision)
{
IRTemp flags0 = newTemp(Ity_I32);
IRTemp flags1 = newTemp(Ity_I32);
IRTemp flags2 = newTemp(Ity_I32);
IRTemp flags3 = newTemp(Ity_I32);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp fe_flag0, fg_flag0, fe_flag1, fg_flag1;
IRTemp fe_flag2, fg_flag2, fe_flag3, fg_flag3;
fe_flag0 = fg_flag0 = fe_flag1 = fg_flag1 = IRTemp_INVALID;
fe_flag2 = fg_flag2 = fe_flag3 = fg_flag3 = IRTemp_INVALID;
DIP("xvtsqrtsp cr%d,v%d\n", (UInt)crfD, (UInt)XB);
breakV128to4x32( getVSReg( XB ), &b3, &b2, &b1, &b0 );
do_fp_tsqrt(b0, True /* single precision*/, &fe_flag0, &fg_flag0);
do_fp_tsqrt(b1, True /* single precision*/, &fe_flag1, &fg_flag1);
do_fp_tsqrt(b2, True /* single precision*/, &fe_flag2, &fg_flag2);
do_fp_tsqrt(b3, True /* single precision*/, &fe_flag3, &fg_flag3);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flags0,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag0), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag0), mkU8( 1 ) ) ) );
assign( flags1,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag1), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag1), mkU8( 1 ) ) ) );
assign( flags2,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag2), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag2), mkU8( 1 ) ) ) );
assign( flags3,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag3), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag3), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR,
binop( Iop_Or32,
mkexpr( flags0 ),
binop( Iop_Or32,
mkexpr( flags1 ),
binop( Iop_Or32,
mkexpr( flags2 ),
mkexpr( flags3 ) ) ) ),
crfD );
break;
}
case 0x174: // xvtdivsp (VSX Vector Test for software Divide Single-Precision)
{
IRTemp flags0 = newTemp(Ity_I32);
IRTemp flags1 = newTemp(Ity_I32);
IRTemp flags2 = newTemp(Ity_I32);
IRTemp flags3 = newTemp(Ity_I32);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp fe_flag0, fg_flag0, fe_flag1, fg_flag1;
IRTemp fe_flag2, fg_flag2, fe_flag3, fg_flag3;
fe_flag0 = fg_flag0 = fe_flag1 = fg_flag1 = IRTemp_INVALID;
fe_flag2 = fg_flag2 = fe_flag3 = fg_flag3 = IRTemp_INVALID;
DIP("xvtdivsp cr%d,v%d,v%d\n", (UInt)crfD, (UInt)XA, (UInt)XB);
breakV128to4x32( getVSReg( XA ), &a3, &a2, &a1, &a0 );
breakV128to4x32( getVSReg( XB ), &b3, &b2, &b1, &b0 );
_do_fp_tdiv(a0, b0, True /* single precision*/, &fe_flag0, &fg_flag0);
_do_fp_tdiv(a1, b1, True /* single precision*/, &fe_flag1, &fg_flag1);
_do_fp_tdiv(a2, b2, True /* single precision*/, &fe_flag2, &fg_flag2);
_do_fp_tdiv(a3, b3, True /* single precision*/, &fe_flag3, &fg_flag3);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flags0,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag0), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag0), mkU8( 1 ) ) ) );
assign( flags1,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag1), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag1), mkU8( 1 ) ) ) );
assign( flags2,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag2), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag2), mkU8( 1 ) ) ) );
assign( flags3,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag3), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag3), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR,
binop( Iop_Or32,
mkexpr( flags0 ),
binop( Iop_Or32,
mkexpr( flags1 ),
binop( Iop_Or32,
mkexpr( flags2 ),
mkexpr( flags3 ) ) ) ),
crfD );
break;
}
default:
vex_printf( "dis_vxv_sp_arith(ppc)(opc2)\n" );
return False;
}
return True;
}
typedef enum {
PPC_CMP_EQ = 2,
PPC_CMP_GT = 4,
PPC_CMP_GE = 6,
PPC_CMP_LT = 8
} ppc_cmp_t;
/*
This helper function takes as input the IRExpr returned
from a binop( Iop_CmpF64, fpA, fpB), whose result is returned
in IR form. This helper function converts it to PPC form.
Map compare result from IR to PPC
FP cmp result | PPC | IR
--------------------------
UN | 0x1 | 0x45
EQ | 0x2 | 0x40
GT | 0x4 | 0x00
LT | 0x8 | 0x01
condcode = Shl(1, (~(ccIR>>5) & 2)
| ((ccIR ^ (ccIR>>6)) & 1)
*/
static IRTemp
get_fp_cmp_CR_val (IRExpr * ccIR_expr)
{
IRTemp condcode = newTemp( Ity_I32 );
IRTemp ccIR = newTemp( Ity_I32 );
assign(ccIR, ccIR_expr);
assign( condcode,
binop( Iop_Shl32,
mkU32( 1 ),
unop( Iop_32to8,
binop( Iop_Or32,
binop( Iop_And32,
unop( Iop_Not32,
binop( Iop_Shr32,
mkexpr( ccIR ),
mkU8( 5 ) ) ),
mkU32( 2 ) ),
binop( Iop_And32,
binop( Iop_Xor32,
mkexpr( ccIR ),
binop( Iop_Shr32,
mkexpr( ccIR ),
mkU8( 6 ) ) ),
mkU32( 1 ) ) ) ) ) );
return condcode;
}
/*
* Helper function for get_max_min_fp for ascertaining the max or min between two doubles
* following these special rules:
* - The max/min of a QNaN and any value is that value
* (When two QNaNs are being compared, the frA QNaN is the return value.)
* - The max/min of any value and an SNaN is that SNaN converted to a QNaN
* (When two SNaNs are being compared, the frA SNaN is converted to a QNaN.)
*/
static IRExpr * _get_maxmin_fp_NaN(IRTemp frA_I64, IRTemp frB_I64)
{
IRTemp frA_isNaN = newTemp(Ity_I1);
IRTemp frB_isNaN = newTemp(Ity_I1);
IRTemp frA_isSNaN = newTemp(Ity_I1);
IRTemp frB_isSNaN = newTemp(Ity_I1);
IRTemp frA_isQNaN = newTemp(Ity_I1);
IRTemp frB_isQNaN = newTemp(Ity_I1);
assign( frA_isNaN, is_NaN( frA_I64 ) );
assign( frB_isNaN, is_NaN( frB_I64 ) );
// If operand is a NAN and bit 12 is '0', then it's an SNaN
assign( frA_isSNaN,
mkAND1( mkexpr(frA_isNaN),
binop( Iop_CmpEQ32,
binop( Iop_And32,
unop( Iop_64HIto32, mkexpr( frA_I64 ) ),
mkU32( 0x00080000 ) ),
mkU32( 0 ) ) ) );
assign( frB_isSNaN,
mkAND1( mkexpr(frB_isNaN),
binop( Iop_CmpEQ32,
binop( Iop_And32,
unop( Iop_64HIto32, mkexpr( frB_I64 ) ),
mkU32( 0x00080000 ) ),
mkU32( 0 ) ) ) );
assign( frA_isQNaN,
mkAND1( mkexpr( frA_isNaN ), unop( Iop_Not1, mkexpr( frA_isSNaN ) ) ) );
assign( frB_isQNaN,
mkAND1( mkexpr( frB_isNaN ), unop( Iop_Not1, mkexpr( frB_isSNaN ) ) ) );
/* Based on the rules specified in the function prologue, the algorithm is as follows:
* <<<<<<<<<>>>>>>>>>>>>>>>>>>
* if frA is a SNaN
* result = frA converted to QNaN
* else if frB is a SNaN
* result = frB converted to QNaN
* else if frB is a QNaN
* result = frA
* // One of frA or frB was a NaN in order for this function to be called, so
* // if we get to this point, we KNOW that frA must be a QNaN.
* else // frA is a QNaN
* result = frB
* <<<<<<<<<>>>>>>>>>>>>>>>>>>
*/
#define SNAN_MASK 0x0008000000000000ULL
return
IRExpr_Mux0X(unop(Iop_1Uto8, mkexpr(frA_isSNaN)),
/* else: if frB is a SNaN */
IRExpr_Mux0X(unop(Iop_1Uto8, mkexpr(frB_isSNaN)),
/* else: if frB is a QNaN */
IRExpr_Mux0X(unop(Iop_1Uto8, mkexpr(frB_isQNaN)),
/* else: frA is a QNaN, so result = frB */
mkexpr(frB_I64),
/* then: result = frA */
mkexpr(frA_I64)),
/* then: result = frB converted to QNaN */
binop(Iop_Or64, mkexpr(frB_I64), mkU64(SNAN_MASK))),
/* then: result = frA converted to QNaN */
binop(Iop_Or64, mkexpr(frA_I64), mkU64(SNAN_MASK)));
}
/*
* Helper function for get_max_min_fp.
*/
static IRExpr * _get_maxmin_fp_cmp(IRTemp src1, IRTemp src2, Bool isMin)
{
IRTemp src1cmpsrc2 = get_fp_cmp_CR_val( binop( Iop_CmpF64,
unop( Iop_ReinterpI64asF64,
mkexpr( src1 ) ),
unop( Iop_ReinterpI64asF64,
mkexpr( src2 ) ) ) );
return IRExpr_Mux0X( unop( Iop_1Uto8,
binop( Iop_CmpEQ32,
mkexpr( src1cmpsrc2 ),
mkU32( isMin ? PPC_CMP_LT : PPC_CMP_GT ) ) ),
/* else: use src2 */
mkexpr( src2 ),
/* then: use src1 */
mkexpr( src1 ) );
}
/*
* Helper function for "Maximum/Minimum Double Precision" operations.
* Arguments: frA and frb are Ity_I64
* Returns Ity_I64 IRExpr that answers the "which is Maxiumum/Minimum" question
*/
static IRExpr * get_max_min_fp(IRTemp frA_I64, IRTemp frB_I64, Bool isMin)
{
/* There are three special cases where get_fp_cmp_CR_val is not helpful
* for ascertaining the maximum between two doubles:
* 1. The max/min of +0 and -0 is +0.
* 2. The max/min of a QNaN and any value is that value.
* 3. The max/min of any value and an SNaN is that SNaN converted to a QNaN.
* We perform the check for [+/-]0 here in this function and use the
* _get_maxmin_fp_NaN helper for the two NaN cases; otherwise we call _get_maxmin_fp_cmp
* to do the standard comparison function.
*/
IRTemp anyNaN = newTemp(Ity_I1);
IRTemp frA_isZero = newTemp(Ity_I1);
IRTemp frB_isZero = newTemp(Ity_I1);
assign(frA_isZero, is_Zero(frA_I64, False /*not single precision*/ ));
assign(frB_isZero, is_Zero(frB_I64, False /*not single precision*/ ));
assign(anyNaN, mkOR1(is_NaN(frA_I64), is_NaN(frB_I64)));
#define MINUS_ZERO 0x8000000000000000ULL
return IRExpr_Mux0X( unop( Iop_1Uto8,
/* If both arguments are zero . . . */
mkAND1( mkexpr( frA_isZero ), mkexpr( frB_isZero ) ) ),
/* else: check if either input is a NaN*/
IRExpr_Mux0X( unop( Iop_1Uto8, mkexpr( anyNaN ) ),
/* else: use "comparison helper" */
_get_maxmin_fp_cmp( frB_I64, frA_I64, isMin ),
/* then: use "NaN helper" */
_get_maxmin_fp_NaN( frA_I64, frB_I64 ) ),
/* then: if frA is -0 and isMin==True, return -0;
* else if frA is +0 and isMin==False; return +0;
* otherwise, simply return frB. */
IRExpr_Mux0X( unop( Iop_1Uto8,
binop( Iop_CmpEQ32,
unop( Iop_64HIto32,
mkexpr( frA_I64 ) ),
mkU32( isMin ? 0x80000000 : 0 ) ) ),
mkexpr( frB_I64 ),
mkU64( isMin ? MINUS_ZERO : 0ULL ) ) );
}
/*
* Helper function for vector/scalar double precision fp round to integer instructions.
*/
static IRExpr * _do_vsx_fp_roundToInt(IRTemp frB_I64, UInt opc2, UChar * insn_suffix)
{
/* The same rules apply for x{s|v}rdpi{m|p|c|z} as for floating point round operations (fri{m|n|p|z}). */
IRTemp frB = newTemp(Ity_F64);
IRTemp frD = newTemp(Ity_F64);
IRTemp intermediateResult = newTemp(Ity_I64);
IRTemp is_SNAN = newTemp(Ity_I1);
IRExpr * hi32;
IRExpr * rxpi_rm;
switch (opc2 & 0x7F) {
case 0x72:
insn_suffix = "m";
rxpi_rm = mkU32(Irrm_NegINF);
break;
case 0x52:
insn_suffix = "p";
rxpi_rm = mkU32(Irrm_PosINF);
break;
case 0x56:
insn_suffix = "c";
rxpi_rm = get_IR_roundingmode();
break;
case 0x32:
insn_suffix = "z";
rxpi_rm = mkU32(Irrm_ZERO);
break;
case 0x12:
insn_suffix = "";
rxpi_rm = mkU32(Irrm_NEAREST);
break;
default: // Impossible to get here
vex_printf( "_do_vsx_fp_roundToInt(ppc)(opc2)\n" );
return NULL;
}
assign(frB, unop(Iop_ReinterpI64asF64, mkexpr(frB_I64)));
assign( intermediateResult,
binop( Iop_F64toI64S, rxpi_rm,
mkexpr( frB ) ) );
/* don't use the rounded integer if frB is outside -9e18..9e18 */
/* F64 has only log10(2**52) significant digits anyway */
/* need to preserve sign of zero */
/* frD = (fabs(frB) > 9e18) ? frB :
(sign(frB)) ? -fabs((double)intermediateResult) : (double)intermediateResult */
assign( frD,
IRExpr_Mux0X( unop( Iop_32to8,
binop( Iop_CmpF64,
IRExpr_Const( IRConst_F64( 9e18 ) ),
unop( Iop_AbsF64, mkexpr( frB ) ) ) ),
IRExpr_Mux0X( unop( Iop_32to8,
binop( Iop_Shr32,
unop( Iop_64HIto32,
mkexpr( frB_I64 ) ),
mkU8( 31 ) ) ),
binop( Iop_I64StoF64,
mkU32( 0 ),
mkexpr( intermediateResult ) ),
unop( Iop_NegF64,
unop( Iop_AbsF64,
binop( Iop_I64StoF64,
mkU32( 0 ),
mkexpr( intermediateResult ) ) ) ) ),
mkexpr( frB ) ) );
/* See Appendix "Floating-Point Round to Integer Model" in ISA doc.
* If frB is a SNAN, then frD <- frB, with bit 12 set to '1'.
*/
#define SNAN_MASK 0x0008000000000000ULL
hi32 = unop( Iop_64HIto32, mkexpr(frB_I64) );
assign( is_SNAN,
mkAND1( is_NaN( frB_I64 ),
binop( Iop_CmpEQ32,
binop( Iop_And32, hi32, mkU32( 0x00080000 ) ),
mkU32( 0 ) ) ) );
return IRExpr_Mux0X( unop( Iop_1Uto8, mkexpr( is_SNAN ) ),
mkexpr( frD ),
unop( Iop_ReinterpI64asF64,
binop( Iop_Xor64,
mkU64( SNAN_MASK ),
mkexpr( frB_I64 ) ) ) );
}
/*
* Miscellaneous VSX vector instructions
*/
static Bool
dis_vxv_misc ( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT( theInstr );
UChar XB = ifieldRegXB( theInstr );
if (opc1 != 0x3C) {
vex_printf( "dis_vxv_misc(ppc)(instr)\n" );
return False;
}
switch (opc2) {
case 0x1B4: // xvredp (VSX Vector Reciprocal Estimate Double-Precision)
case 0x194: // xvrsqrtedp (VSX Vector Reciprocal Square Root Estimate
// Double-Precision)
{
IRExpr* ieee_one = IRExpr_Const(IRConst_F64i(0x3ff0000000000000ULL));
IRExpr* rm = get_IR_roundingmode();
IRTemp frB = newTemp(Ity_I64);
IRTemp frB2 = newTemp(Ity_I64);
Bool redp = opc2 == 0x1B4;
IRTemp sqrtHi = newTemp(Ity_F64);
IRTemp sqrtLo = newTemp(Ity_F64);
assign(frB, unop(Iop_V128HIto64, getVSReg( XB )));
assign(frB2, unop(Iop_V128to64, getVSReg( XB )));
DIP("%s v%d,v%d\n", redp ? "xvredp" : "xvrsqrtedp", (UInt)XT, (UInt)XB);
if (!redp) {
assign( sqrtHi,
binop( Iop_SqrtF64,
rm,
unop( Iop_ReinterpI64asF64, mkexpr( frB ) ) ) );
assign( sqrtLo,
binop( Iop_SqrtF64,
rm,
unop( Iop_ReinterpI64asF64, mkexpr( frB2 ) ) ) );
}
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
triop( Iop_DivF64,
rm,
ieee_one,
redp ? unop( Iop_ReinterpI64asF64,
mkexpr( frB ) )
: mkexpr( sqrtHi ) ) ),
unop( Iop_ReinterpF64asI64,
triop( Iop_DivF64,
rm,
ieee_one,
redp ? unop( Iop_ReinterpI64asF64,
mkexpr( frB2 ) )
: mkexpr( sqrtLo ) ) ) ) );
break;
}
case 0x134: // xvresp (VSX Vector Reciprocal Estimate Single-Precision)
case 0x114: // xvrsqrtesp (VSX Vector Reciprocal Square Root Estimate Single-Precision)
{
IRTemp b3, b2, b1, b0;
IRTemp res0 = newTemp(Ity_I32);
IRTemp res1 = newTemp(Ity_I32);
IRTemp res2 = newTemp(Ity_I32);
IRTemp res3 = newTemp(Ity_I32);
IRTemp sqrt3 = newTemp(Ity_F64);
IRTemp sqrt2 = newTemp(Ity_F64);
IRTemp sqrt1 = newTemp(Ity_F64);
IRTemp sqrt0 = newTemp(Ity_F64);
IRExpr* rm = get_IR_roundingmode();
Bool resp = opc2 == 0x134;
IRExpr* ieee_one = IRExpr_Const(IRConst_F64i(0x3ff0000000000000ULL));
b3 = b2 = b1 = b0 = IRTemp_INVALID;
DIP("%s v%d,v%d\n", resp ? "xvresp" : "xvrsqrtesp", (UInt)XT, (UInt)XB);
breakV128to4xF64( getVSReg( XB ), &b3, &b2, &b1, &b0 );
if (!resp) {
assign( sqrt3, binop( Iop_SqrtF64, rm, mkexpr( b3 ) ) );
assign( sqrt2, binop( Iop_SqrtF64, rm, mkexpr( b2 ) ) );
assign( sqrt1, binop( Iop_SqrtF64, rm, mkexpr( b1 ) ) );
assign( sqrt0, binop( Iop_SqrtF64, rm, mkexpr( b0 ) ) );
}
assign( res0,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32,
rm,
ieee_one,
resp ? mkexpr( b0 ) : mkexpr( sqrt0 ) ) ) ) );
assign( res1,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32,
rm,
ieee_one,
resp ? mkexpr( b1 ) : mkexpr( sqrt1 ) ) ) ) );
assign( res2,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32,
rm,
ieee_one,
resp ? mkexpr( b2 ) : mkexpr( sqrt2 ) ) ) ) );
assign( res3,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
triop( Iop_DivF64r32,
rm,
ieee_one,
resp ? mkexpr( b3 ) : mkexpr( sqrt3 ) ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkexpr( res3 ), mkexpr( res2 ) ),
binop( Iop_32HLto64, mkexpr( res1 ), mkexpr( res0 ) ) ) );
break;
}
case 0x300: // xvmaxsp (VSX Vector Maximum Single-Precision)
case 0x320: // xvminsp (VSX Vector Minimum Single-Precision)
{
UChar XA = ifieldRegXA( theInstr );
IRTemp a3, a2, a1, a0;
IRTemp b3, b2, b1, b0;
IRTemp res0 = newTemp( Ity_I32 );
IRTemp res1 = newTemp( Ity_I32 );
IRTemp res2 = newTemp( Ity_I32 );
IRTemp res3 = newTemp( Ity_I32 );
IRTemp a0_I64 = newTemp( Ity_I64 );
IRTemp a1_I64 = newTemp( Ity_I64 );
IRTemp a2_I64 = newTemp( Ity_I64 );
IRTemp a3_I64 = newTemp( Ity_I64 );
IRTemp b0_I64 = newTemp( Ity_I64 );
IRTemp b1_I64 = newTemp( Ity_I64 );
IRTemp b2_I64 = newTemp( Ity_I64 );
IRTemp b3_I64 = newTemp( Ity_I64 );
Bool isMin = opc2 == 0x320 ? True : False;
a3 = a2 = a1 = a0 = IRTemp_INVALID;
b3 = b2 = b1 = b0 = IRTemp_INVALID;
DIP("%s v%d,v%d v%d\n", isMin ? "xvminsp" : "xvmaxsp", (UInt)XT, (UInt)XA, (UInt)XB);
breakV128to4xF64( getVSReg( XA ), &a3, &a2, &a1, &a0 );
breakV128to4xF64( getVSReg( XB ), &b3, &b2, &b1, &b0 );
assign( a0_I64, unop( Iop_ReinterpF64asI64, mkexpr( a0 ) ) );
assign( b0_I64, unop( Iop_ReinterpF64asI64, mkexpr( b0 ) ) );
assign( a1_I64, unop( Iop_ReinterpF64asI64, mkexpr( a1 ) ) );
assign( b1_I64, unop( Iop_ReinterpF64asI64, mkexpr( b1 ) ) );
assign( a2_I64, unop( Iop_ReinterpF64asI64, mkexpr( a2 ) ) );
assign( b2_I64, unop( Iop_ReinterpF64asI64, mkexpr( b2 ) ) );
assign( a3_I64, unop( Iop_ReinterpF64asI64, mkexpr( a3 ) ) );
assign( b3_I64, unop( Iop_ReinterpF64asI64, mkexpr( b3 ) ) );
assign( res0,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
unop( Iop_ReinterpI64asF64,
get_max_min_fp( a0_I64, b0_I64, isMin ) ) ) ) );
assign( res1,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
unop( Iop_ReinterpI64asF64,
get_max_min_fp( a1_I64, b1_I64, isMin ) ) ) ) );
assign( res2,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
unop( Iop_ReinterpI64asF64,
get_max_min_fp( a2_I64, b2_I64, isMin ) ) ) ) );
assign( res3,
unop( Iop_ReinterpF32asI32,
unop( Iop_TruncF64asF32,
unop( Iop_ReinterpI64asF64,
get_max_min_fp( a3_I64, b3_I64, isMin ) ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkexpr( res3 ), mkexpr( res2 ) ),
binop( Iop_32HLto64, mkexpr( res1 ), mkexpr( res0 ) ) ) );
break;
}
case 0x380: // xvmaxdp (VSX Vector Maximum Double-Precision)
case 0x3A0: // xvmindp (VSX Vector Minimum Double-Precision)
{
UChar XA = ifieldRegXA( theInstr );
IRTemp frA = newTemp(Ity_I64);
IRTemp frB = newTemp(Ity_I64);
IRTemp frA2 = newTemp(Ity_I64);
IRTemp frB2 = newTemp(Ity_I64);
Bool isMin = opc2 == 0x3A0 ? True : False;
assign(frA, unop(Iop_V128HIto64, getVSReg( XA )));
assign(frB, unop(Iop_V128HIto64, getVSReg( XB )));
assign(frA2, unop(Iop_V128to64, getVSReg( XA )));
assign(frB2, unop(Iop_V128to64, getVSReg( XB )));
DIP("%s v%d,v%d v%d\n", isMin ? "xvmindp" : "xvmaxdp", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, get_max_min_fp(frA, frB, isMin), get_max_min_fp(frA2, frB2, isMin) ) );
break;
}
case 0x3c0: // xvcpsgndp (VSX Vector Copy Sign Double-Precision)
{
UChar XA = ifieldRegXA( theInstr );
IRTemp frA = newTemp(Ity_I64);
IRTemp frB = newTemp(Ity_I64);
IRTemp frA2 = newTemp(Ity_I64);
IRTemp frB2 = newTemp(Ity_I64);
assign(frA, unop(Iop_V128HIto64, getVSReg( XA )));
assign(frB, unop(Iop_V128HIto64, getVSReg( XB )));
assign(frA2, unop(Iop_V128to64, getVSReg( XA )));
assign(frB2, unop(Iop_V128to64, getVSReg( XB )));
DIP("xvcpsgndp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_Or64,
binop( Iop_And64,
mkexpr( frA ),
mkU64( SIGN_BIT ) ),
binop( Iop_And64,
mkexpr( frB ),
mkU64( SIGN_MASK ) ) ),
binop( Iop_Or64,
binop( Iop_And64,
mkexpr( frA2 ),
mkU64( SIGN_BIT ) ),
binop( Iop_And64,
mkexpr( frB2 ),
mkU64( SIGN_MASK ) ) ) ) );
break;
}
case 0x340: // xvcpsgnsp
{
UChar XA = ifieldRegXA( theInstr );
IRTemp a3_I64, a2_I64, a1_I64, a0_I64;
IRTemp b3_I64, b2_I64, b1_I64, b0_I64;
IRTemp resHi = newTemp(Ity_I64);
IRTemp resLo = newTemp(Ity_I64);
a3_I64 = a2_I64 = a1_I64 = a0_I64 = IRTemp_INVALID;
b3_I64 = b2_I64 = b1_I64 = b0_I64 = IRTemp_INVALID;
DIP("xvcpsgnsp v%d,v%d v%d\n",(UInt)XT, (UInt)XA, (UInt)XB);
breakV128to4x64U( getVSReg( XA ), &a3_I64, &a2_I64, &a1_I64, &a0_I64 );
breakV128to4x64U( getVSReg( XB ), &b3_I64, &b2_I64, &b1_I64, &b0_I64 );
assign( resHi,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
unop(Iop_64to32, mkexpr( a3_I64 ) ),
mkU32( SIGN_BIT32 ) ),
binop( Iop_And32,
unop(Iop_64to32, mkexpr( b3_I64 ) ),
mkU32( SIGN_MASK32) ) ),
binop( Iop_Or32,
binop( Iop_And32,
unop(Iop_64to32, mkexpr( a2_I64 ) ),
mkU32( SIGN_BIT32 ) ),
binop( Iop_And32,
unop(Iop_64to32, mkexpr( b2_I64 ) ),
mkU32( SIGN_MASK32 ) ) ) ) );
assign( resLo,
binop( Iop_32HLto64,
binop( Iop_Or32,
binop( Iop_And32,
unop(Iop_64to32, mkexpr( a1_I64 ) ),
mkU32( SIGN_BIT32 ) ),
binop( Iop_And32,
unop(Iop_64to32, mkexpr( b1_I64 ) ),
mkU32( SIGN_MASK32 ) ) ),
binop( Iop_Or32,
binop( Iop_And32,
unop(Iop_64to32, mkexpr( a0_I64 ) ),
mkU32( SIGN_BIT32 ) ),
binop( Iop_And32,
unop(Iop_64to32, mkexpr( b0_I64 ) ),
mkU32( SIGN_MASK32 ) ) ) ) );
putVSReg( XT, binop( Iop_64HLtoV128, mkexpr( resHi ), mkexpr( resLo ) ) );
break;
}
case 0x3B2: // xvabsdp (VSX Vector Absolute Value Double-Precision)
case 0x3D2: // xvnabsdp VSX Vector Negative Absolute Value Double-Precision)
{
IRTemp frB = newTemp(Ity_F64);
IRTemp frB2 = newTemp(Ity_F64);
IRTemp abs_resultHi = newTemp(Ity_F64);
IRTemp abs_resultLo = newTemp(Ity_F64);
Bool make_negative = (opc2 == 0x3D2) ? True : False;
assign(frB, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XB ))));
assign(frB2, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, getVSReg(XB))));
DIP("xv%sabsdp v%d,v%d\n", make_negative ? "n" : "", (UInt)XT, (UInt)XB);
if (make_negative) {
assign(abs_resultHi, unop( Iop_NegF64, unop( Iop_AbsF64, mkexpr( frB ) ) ) );
assign(abs_resultLo, unop( Iop_NegF64, unop( Iop_AbsF64, mkexpr( frB2 ) ) ) );
} else {
assign(abs_resultHi, unop( Iop_AbsF64, mkexpr( frB ) ) );
assign(abs_resultLo, unop( Iop_AbsF64, mkexpr( frB2 ) ) );
}
putVSReg( XT, binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64, mkexpr( abs_resultHi ) ),
unop( Iop_ReinterpF64asI64, mkexpr( abs_resultLo ) ) ) );
break;
}
case 0x332: // xvabssp (VSX Vector Absolute Value Single-Precision)
case 0x352: // xvnabssp (VSX Vector Negative Absolute Value Single-Precision)
{
/*
* The Iop_AbsF32 IRop is not implemented for ppc64 since, up until introduction
* of xvabssp, there has not been an abs(sp) type of instruction. But since emulation
* of this function is so easy using shifts, I choose to emulate this instruction that
* way versus a native instruction method of implementation.
*/
Bool make_negative = (opc2 == 0x352) ? True : False;
IRTemp shiftVector = newTemp(Ity_V128);
IRTemp absVal_vector = newTemp(Ity_V128);
assign( shiftVector,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64, mkU32( 1 ), mkU32( 1 ) ),
binop( Iop_32HLto64, mkU32( 1 ), mkU32( 1 ) ) ) );
assign( absVal_vector,
binop( Iop_Shr32x4,
binop( Iop_Shl32x4,
getVSReg( XB ),
mkexpr( shiftVector ) ),
mkexpr( shiftVector ) ) );
if (make_negative) {
IRTemp signBit_vector = newTemp(Ity_V128);
assign( signBit_vector,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
mkU32( 0x80000000 ),
mkU32( 0x80000000 ) ),
binop( Iop_32HLto64,
mkU32( 0x80000000 ),
mkU32( 0x80000000 ) ) ) );
putVSReg( XT,
binop( Iop_OrV128,
mkexpr( absVal_vector ),
mkexpr( signBit_vector ) ) );
} else {
putVSReg( XT, mkexpr( absVal_vector ) );
}
break;
}
case 0x3F2: // xvnegdp (VSX Vector Negate Double-Precision)
{
IRTemp frB = newTemp(Ity_F64);
IRTemp frB2 = newTemp(Ity_F64);
assign(frB, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XB ))));
assign(frB2, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, getVSReg(XB))));
DIP("xvnegdp v%d,v%d\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
unop( Iop_NegF64, mkexpr( frB ) ) ),
unop( Iop_ReinterpF64asI64,
unop( Iop_NegF64, mkexpr( frB2 ) ) ) ) );
break;
}
case 0x192: // xvrdpi (VSX Vector Round to Double-Precision Integer using round toward Nearest Away)
case 0x1D6: // xvrdpic (VSX Vector Round to Double-Precision Integer using Current rounding mode)
case 0x1F2: // xvrdpim (VSX Vector Round to Double-Precision Integer using round toward -Infinity)
case 0x1D2: // xvrdpip (VSX Vector Round to Double-Precision Integer using round toward +Infinity)
case 0x1B2: // xvrdpiz (VSX Vector Round to Double-Precision Integer using round toward Zero)
{
IRTemp frBHi_I64 = newTemp(Ity_I64);
IRTemp frBLo_I64 = newTemp(Ity_I64);
IRExpr * frD_fp_roundHi = NULL;
IRExpr * frD_fp_roundLo = NULL;
UChar * insn_suffix = NULL;
assign( frBHi_I64, unop( Iop_V128HIto64, getVSReg( XB ) ) );
frD_fp_roundHi = _do_vsx_fp_roundToInt(frBHi_I64, opc2, insn_suffix);
assign( frBLo_I64, unop( Iop_V128to64, getVSReg( XB ) ) );
frD_fp_roundLo = _do_vsx_fp_roundToInt(frBLo_I64, opc2, insn_suffix);
DIP("xvrdpi%s v%d,v%d\n", insn_suffix, (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64, frD_fp_roundHi ),
unop( Iop_ReinterpF64asI64, frD_fp_roundLo ) ) );
break;
}
case 0x112: // xvrspi (VSX Vector Round to Single-Precision Integer using round toward Nearest Away)
case 0x156: // xvrspic (VSX Vector Round to SinglePrecision Integer using Current rounding mode)
case 0x172: // xvrspim (VSX Vector Round to SinglePrecision Integer using round toward -Infinity)
case 0x152: // xvrspip (VSX Vector Round to SinglePrecision Integer using round toward +Infinity)
case 0x132: // xvrspiz (VSX Vector Round to SinglePrecision Integer using round toward Zero)
{
UChar * insn_suffix = NULL;
IROp op;
if (opc2 != 0x156) {
// Use pre-defined IRop's for vrfi{m|n|p|z}
switch (opc2) {
case 0x112:
insn_suffix = "";
op = Iop_RoundF32x4_RN;
break;
case 0x172:
insn_suffix = "m";
op = Iop_RoundF32x4_RM;
break;
case 0x152:
insn_suffix = "p";
op = Iop_RoundF32x4_RP;
break;
case 0x132:
insn_suffix = "z";
op = Iop_RoundF32x4_RZ;
break;
default:
vex_printf( "dis_vxv_misc(ppc)(vrspi<x>)(opc2)\n" );
return False;
}
DIP("xvrspi%s v%d,v%d\n", insn_suffix, (UInt)XT, (UInt)XB);
putVSReg( XT, unop( op, getVSReg(XB) ) );
} else {
// Handle xvrspic. Unfortunately there is no corresponding "vfric" instruction.
IRExpr * frD_fp_roundb3, * frD_fp_roundb2, * frD_fp_roundb1, * frD_fp_roundb0;
IRTemp b3_F64, b2_F64, b1_F64, b0_F64;
IRTemp b3_I64 = newTemp(Ity_I64);
IRTemp b2_I64 = newTemp(Ity_I64);
IRTemp b1_I64 = newTemp(Ity_I64);
IRTemp b0_I64 = newTemp(Ity_I64);
b3_F64 = b2_F64 = b1_F64 = b0_F64 = IRTemp_INVALID;
frD_fp_roundb3 = frD_fp_roundb2 = frD_fp_roundb1 = frD_fp_roundb0 = NULL;
breakV128to4xF64( getVSReg(XB), &b3_F64, &b2_F64, &b1_F64, &b0_F64);
assign(b3_I64, unop(Iop_ReinterpF64asI64, mkexpr(b3_F64)));
assign(b2_I64, unop(Iop_ReinterpF64asI64, mkexpr(b2_F64)));
assign(b1_I64, unop(Iop_ReinterpF64asI64, mkexpr(b1_F64)));
assign(b0_I64, unop(Iop_ReinterpF64asI64, mkexpr(b0_F64)));
frD_fp_roundb3 = unop(Iop_TruncF64asF32,
_do_vsx_fp_roundToInt(b3_I64, opc2, insn_suffix));
frD_fp_roundb2 = unop(Iop_TruncF64asF32,
_do_vsx_fp_roundToInt(b2_I64, opc2, insn_suffix));
frD_fp_roundb1 = unop(Iop_TruncF64asF32,
_do_vsx_fp_roundToInt(b1_I64, opc2, insn_suffix));
frD_fp_roundb0 = unop(Iop_TruncF64asF32,
_do_vsx_fp_roundToInt(b0_I64, opc2, insn_suffix));
DIP("xvrspic v%d,v%d\n", (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32, frD_fp_roundb3 ),
unop( Iop_ReinterpF32asI32, frD_fp_roundb2 ) ),
binop( Iop_32HLto64,
unop( Iop_ReinterpF32asI32, frD_fp_roundb1 ),
unop( Iop_ReinterpF32asI32, frD_fp_roundb0 ) ) ) );
}
break;
}
default:
vex_printf( "dis_vxv_misc(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX Scalar Floating Point Arithmetic Instructions
*/
static Bool
dis_vxs_arith ( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT( theInstr );
UChar XA = ifieldRegXA( theInstr );
UChar XB = ifieldRegXB( theInstr );
IRExpr* rm = get_IR_roundingmode();
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
if (opc1 != 0x3C) {
vex_printf( "dis_vxs_arith(ppc)(instr)\n" );
return False;
}
assign(frA, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XA ))));
assign(frB, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XB ))));
/* For all the VSX sclar arithmetic instructions, the contents of doubleword element 1
* of VSX[XT] are undefined after the operation; therefore, we can simply set
* element to zero where it makes sense to do so.
*/
switch (opc2) {
case 0x080: // xsadddp (VSX scalar add double-precision)
DIP("xsadddp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
triop( Iop_AddF64, rm,
mkexpr( frA ),
mkexpr( frB ) ) ),
mkU64( 0 ) ) );
break;
case 0x0E0: // xsdivdp (VSX scalar divide double-precision)
DIP("xsdivdp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
triop( Iop_DivF64, rm,
mkexpr( frA ),
mkexpr( frB ) ) ),
mkU64( 0 ) ) );
break;
case 0x084: case 0x0A4: // xsmaddadp, xsmaddmdp (VSX scalar multiply-add double-precision)
{
IRTemp frT = newTemp(Ity_F64);
Bool mdp = opc2 == 0x0A4;
DIP("xsmadd%sdp v%d,v%d,v%d\n", mdp ? "m" : "a", (UInt)XT, (UInt)XA, (UInt)XB);
assign( frT, unop( Iop_ReinterpI64asF64, unop( Iop_V128HIto64,
getVSReg( XT ) ) ) );
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
qop( Iop_MAddF64, rm,
mkexpr( frA ),
mkexpr( mdp ? frT : frB ),
mkexpr( mdp ? frB : frT ) ) ),
mkU64( 0 ) ) );
break;
}
case 0x0C4: case 0x0E4: // xsmsubadp, xsmsubmdp (VSX scalar multiply-subtract double-precision)
{
IRTemp frT = newTemp(Ity_F64);
Bool mdp = opc2 == 0x0E4;
DIP("xsmsub%sdp v%d,v%d,v%d\n", mdp ? "m" : "a", (UInt)XT, (UInt)XA, (UInt)XB);
assign( frT, unop( Iop_ReinterpI64asF64, unop( Iop_V128HIto64,
getVSReg( XT ) ) ) );
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
qop( Iop_MSubF64, rm,
mkexpr( frA ),
mkexpr( mdp ? frT : frB ),
mkexpr( mdp ? frB : frT ) ) ),
mkU64( 0 ) ) );
break;
}
case 0x284: case 0x2A4: // xsnmaddadp, xsnmaddmdp (VSX scalar multiply-add double-precision)
{
/* TODO: mpj -- Naturally, I expected to be able to leverage the implementation
* of fnmadd and use pretty much the same code. However, that code has a bug in the
* way it blindly negates the signbit, even if the floating point result is a NaN.
* So, the TODO is to fix fnmadd (which I'll do in a different patch).
*/
Bool mdp = opc2 == 0x2A4;
IRTemp frT = newTemp(Ity_F64);
IRTemp maddResult = newTemp(Ity_I64);
DIP("xsnmadd%sdp v%d,v%d,v%d\n", mdp ? "m" : "a", (UInt)XT, (UInt)XA, (UInt)XB);
assign( frT, unop( Iop_ReinterpI64asF64, unop( Iop_V128HIto64,
getVSReg( XT ) ) ) );
assign( maddResult, unop( Iop_ReinterpF64asI64, qop( Iop_MAddF64, rm,
mkexpr( frA ),
mkexpr( mdp ? frT : frB ),
mkexpr( mdp ? frB : frT ) ) ) );
putVSReg( XT, binop( Iop_64HLtoV128, mkexpr( getNegatedResult(maddResult) ),
mkU64( 0 ) ) );
break;
}
case 0x2C4: case 0x2E4: // xsnmsubadp, xsnmsubmdp (VSX Scalar Negative Multiply-Subtract Double-Precision)
{
IRTemp frT = newTemp(Ity_F64);
Bool mdp = opc2 == 0x2E4;
IRTemp msubResult = newTemp(Ity_I64);
DIP("xsnmsub%sdp v%d,v%d,v%d\n", mdp ? "m" : "a", (UInt)XT, (UInt)XA, (UInt)XB);
assign( frT, unop( Iop_ReinterpI64asF64, unop( Iop_V128HIto64,
getVSReg( XT ) ) ) );
assign(msubResult, unop( Iop_ReinterpF64asI64,
qop( Iop_MSubF64,
rm,
mkexpr( frA ),
mkexpr( mdp ? frT : frB ),
mkexpr( mdp ? frB : frT ) ) ));
putVSReg( XT, binop( Iop_64HLtoV128, mkexpr( getNegatedResult(msubResult) ), mkU64( 0 ) ) );
break;
}
case 0x0C0: // xsmuldp (VSX Scalar Multiply Double-Precision)
DIP("xsmuldp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
triop( Iop_MulF64, rm,
mkexpr( frA ),
mkexpr( frB ) ) ),
mkU64( 0 ) ) );
break;
case 0x0A0: // xssubdp (VSX Scalar Subtract Double-Precision)
DIP("xssubdp v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
triop( Iop_SubF64, rm,
mkexpr( frA ),
mkexpr( frB ) ) ),
mkU64( 0 ) ) );
break;
case 0x096: // xssqrtdp (VSX Scalar Square Root Double-Precision)
DIP("xssqrtdp v%d,v%d\n", (UInt)XT, (UInt)XB);
putVSReg( XT, binop( Iop_64HLtoV128, unop( Iop_ReinterpF64asI64,
binop( Iop_SqrtF64, rm,
mkexpr( frB ) ) ),
mkU64( 0 ) ) );
break;
case 0x0F4: // xstdivdp (VSX Scalar Test for software Divide Double-Precision)
{
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp frA_I64 = newTemp(Ity_I64);
IRTemp frB_I64 = newTemp(Ity_I64);
DIP("xstdivdp crf%d,v%d,v%d\n", crfD, (UInt)XA, (UInt)XB);
assign( frA_I64, unop( Iop_ReinterpF64asI64, mkexpr( frA ) ) );
assign( frB_I64, unop( Iop_ReinterpF64asI64, mkexpr( frB ) ) );
putGST_field( PPC_GST_CR, do_fp_tdiv(frA_I64, frB_I64), crfD );
break;
}
case 0x0D4: // xstsqrtdp (VSX Vector Test for software Square Root Double-Precision)
{
IRTemp frB_I64 = newTemp(Ity_I64);
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp flags = newTemp(Ity_I32);
IRTemp fe_flag, fg_flag;
fe_flag = fg_flag = IRTemp_INVALID;
DIP("xstsqrtdp v%d,v%d\n", (UInt)XT, (UInt)XB);
assign( frB_I64, unop(Iop_V128HIto64, getVSReg( XB )) );
do_fp_tsqrt(frB_I64, False /*not single precision*/, &fe_flag, &fg_flag);
/* The CR field consists of fl_flag || fg_flag || fe_flag || 0b0
* where fl_flag == 1 on ppc64.
*/
assign( flags,
binop( Iop_Or32,
binop( Iop_Or32, mkU32( 8 ), // fl_flag
binop( Iop_Shl32, mkexpr(fg_flag), mkU8( 2 ) ) ),
binop( Iop_Shl32, mkexpr(fe_flag), mkU8( 1 ) ) ) );
putGST_field( PPC_GST_CR, mkexpr(flags), crfD );
break;
}
default:
vex_printf( "dis_vxs_arith(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX Floating Point Compare Instructions
*/
static Bool
dis_vx_cmp( UInt theInstr, UInt opc2 )
{
/* XX3-Form and XX2-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar crfD = toUChar( IFIELD( theInstr, 23, 3 ) );
IRTemp ccPPC32;
UChar XA = ifieldRegXA ( theInstr );
UChar XB = ifieldRegXB ( theInstr );
IRTemp frA = newTemp(Ity_F64);
IRTemp frB = newTemp(Ity_F64);
if (opc1 != 0x3C) {
vex_printf( "dis_vx_cmp(ppc)(instr)\n" );
return False;
}
assign(frA, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XA ))));
assign(frB, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, getVSReg( XB ))));
switch (opc2) {
case 0x08C: case 0x0AC: // xscmpudp, xscmpodp
/* Note: Differences between xscmpudp and xscmpodp are only in
* exception flag settings, which aren't supported anyway. */
DIP("xscmp%sdp crf%d,fr%u,fr%u\n", opc2 == 0x08c ? "u" : "o",
crfD, (UInt)XA, (UInt)XB);
ccPPC32 = get_fp_cmp_CR_val( binop(Iop_CmpF64, mkexpr(frA), mkexpr(frB)));
putGST_field( PPC_GST_CR, mkexpr(ccPPC32), crfD );
break;
default:
vex_printf( "dis_vx_cmp(ppc)(opc2)\n" );
return False;
}
return True;
}
static void
do_vvec_fp_cmp ( IRTemp vA, IRTemp vB, UChar XT, UChar flag_rC,
ppc_cmp_t cmp_type )
{
IRTemp frA_hi = newTemp(Ity_F64);
IRTemp frB_hi = newTemp(Ity_F64);
IRTemp frA_lo = newTemp(Ity_F64);
IRTemp frB_lo = newTemp(Ity_F64);
IRTemp ccPPC32 = newTemp(Ity_I32);
IRTemp ccIR_hi;
IRTemp ccIR_lo;
IRTemp hiResult = newTemp(Ity_I64);
IRTemp loResult = newTemp(Ity_I64);
IRTemp hiEQlo = newTemp(Ity_I1);
IRTemp all_elem_true = newTemp(Ity_I32);
IRTemp all_elem_false = newTemp(Ity_I32);
assign(frA_hi, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, mkexpr( vA ))));
assign(frB_hi, unop(Iop_ReinterpI64asF64, unop(Iop_V128HIto64, mkexpr( vB ))));
assign(frA_lo, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, mkexpr( vA ))));
assign(frB_lo, unop(Iop_ReinterpI64asF64, unop(Iop_V128to64, mkexpr( vB ))));
ccIR_hi = get_fp_cmp_CR_val( binop( Iop_CmpF64,
mkexpr( frA_hi ),
mkexpr( frB_hi ) ) );
ccIR_lo = get_fp_cmp_CR_val( binop( Iop_CmpF64,
mkexpr( frA_lo ),
mkexpr( frB_lo ) ) );
if (cmp_type != PPC_CMP_GE) {
assign( hiResult,
unop( Iop_1Sto64,
binop( Iop_CmpEQ32, mkexpr( ccIR_hi ), mkU32( cmp_type ) ) ) );
assign( loResult,
unop( Iop_1Sto64,
binop( Iop_CmpEQ32, mkexpr( ccIR_lo ), mkU32( cmp_type ) ) ) );
} else {
// For PPC_CMP_GE, one element compare may return "4" (for "greater than") and
// the other element compare may return "2" (for "equal to").
IRTemp lo_GE = newTemp(Ity_I1);
IRTemp hi_GE = newTemp(Ity_I1);
assign(hi_GE, mkOR1( binop( Iop_CmpEQ32, mkexpr( ccIR_hi ), mkU32( 2 ) ),
binop( Iop_CmpEQ32, mkexpr( ccIR_hi ), mkU32( 4 ) ) ) );
assign( hiResult,unop( Iop_1Sto64, mkexpr( hi_GE ) ) );
assign(lo_GE, mkOR1( binop( Iop_CmpEQ32, mkexpr( ccIR_lo ), mkU32( 2 ) ),
binop( Iop_CmpEQ32, mkexpr( ccIR_lo ), mkU32( 4 ) ) ) );
assign( loResult, unop( Iop_1Sto64, mkexpr( lo_GE ) ) );
}
// The [hi/lo]Result will be all 1's or all 0's. We just look at the lower word.
assign( hiEQlo,
binop( Iop_CmpEQ32,
unop( Iop_64to32, mkexpr( hiResult ) ),
unop( Iop_64to32, mkexpr( loResult ) ) ) );
putVSReg( XT,
binop( Iop_64HLtoV128, mkexpr( hiResult ), mkexpr( loResult ) ) );
assign( all_elem_true,
unop( Iop_1Uto32,
mkAND1( mkexpr( hiEQlo ),
binop( Iop_CmpEQ32,
mkU32( 0xffffffff ),
unop( Iop_64to32,
mkexpr( hiResult ) ) ) ) ) );
assign( all_elem_false,
unop( Iop_1Uto32,
mkAND1( mkexpr( hiEQlo ),
binop( Iop_CmpEQ32,
mkU32( 0 ),
unop( Iop_64to32,
mkexpr( hiResult ) ) ) ) ) );
assign( ccPPC32,
binop( Iop_Or32,
binop( Iop_Shl32, mkexpr( all_elem_false ), mkU8( 1 ) ),
binop( Iop_Shl32, mkexpr( all_elem_true ), mkU8( 3 ) ) ) );
if (flag_rC) {
putGST_field( PPC_GST_CR, mkexpr(ccPPC32), 6 );
}
}
/*
* VSX Vector Compare Instructions
*/
static Bool
dis_vvec_cmp( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT ( theInstr );
UChar XA = ifieldRegXA ( theInstr );
UChar XB = ifieldRegXB ( theInstr );
UChar flag_rC = ifieldBIT10(theInstr);
IRTemp vA = newTemp( Ity_V128 );
IRTemp vB = newTemp( Ity_V128 );
if (opc1 != 0x3C) {
vex_printf( "dis_vvec_cmp(ppc)(instr)\n" );
return False;
}
assign( vA, getVSReg( XA ) );
assign( vB, getVSReg( XB ) );
switch (opc2) {
case 0x18C: case 0x38C: // xvcmpeqdp[.] (VSX Vector Compare Equal To Double-Precision [ & Record ])
{
DIP("xvcmpeqdp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
do_vvec_fp_cmp(vA, vB, XT, flag_rC, PPC_CMP_EQ);
break;
}
case 0x1CC: case 0x3CC: // xvcmpgedp[.] (VSX Vector Compare Greater Than or Equal To Double-Precision [ & Record ])
{
DIP("xvcmpgedp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
do_vvec_fp_cmp(vA, vB, XT, flag_rC, PPC_CMP_GE);
break;
}
case 0x1AC: case 0x3AC: // xvcmpgtdp[.] (VSX Vector Compare Greater Than Double-Precision [ & Record ])
{
DIP("xvcmpgtdp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
do_vvec_fp_cmp(vA, vB, XT, flag_rC, PPC_CMP_GT);
break;
}
case 0x10C: case 0x30C: // xvcmpeqsp[.] (VSX Vector Compare Equal To Single-Precision [ & Record ])
{
IRTemp vD = newTemp(Ity_V128);
DIP("xvcmpeqsp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
assign( vD, binop(Iop_CmpEQ32Fx4, mkexpr(vA), mkexpr(vB)) );
putVSReg( XT, mkexpr(vD) );
if (flag_rC) {
set_AV_CR6( mkexpr(vD), True );
}
break;
}
case 0x14C: case 0x34C: // xvcmpgesp[.] (VSX Vector Compare Greater Than or Equal To Single-Precision [ & Record ])
{
IRTemp vD = newTemp(Ity_V128);
DIP("xvcmpgesp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
assign( vD, binop(Iop_CmpGE32Fx4, mkexpr(vA), mkexpr(vB)) );
putVSReg( XT, mkexpr(vD) );
if (flag_rC) {
set_AV_CR6( mkexpr(vD), True );
}
break;
}
case 0x12C: case 0x32C: //xvcmpgtsp[.] (VSX Vector Compare Greater Than Single-Precision [ & Record ])
{
IRTemp vD = newTemp(Ity_V128);
DIP("xvcmpgtsp%s crf%d,fr%u,fr%u\n", (flag_rC ? ".":""),
(UInt)XT, (UInt)XA, (UInt)XB);
assign( vD, binop(Iop_CmpGT32Fx4, mkexpr(vA), mkexpr(vB)) );
putVSReg( XT, mkexpr(vD) );
if (flag_rC) {
set_AV_CR6( mkexpr(vD), True );
}
break;
}
default:
vex_printf( "dis_vvec_cmp(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* Miscellaneous VSX Scalar Instructions
*/
static Bool
dis_vxs_misc( UInt theInstr, UInt opc2 )
{
/* XX3-Form and XX2-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT ( theInstr );
UChar XA = ifieldRegXA ( theInstr );
UChar XB = ifieldRegXB ( theInstr );
IRTemp vA = newTemp( Ity_V128 );
IRTemp vB = newTemp( Ity_V128 );
if (opc1 != 0x3C) {
vex_printf( "dis_vxs_misc(ppc)(instr)\n" );
return False;
}
assign( vA, getVSReg( XA ) );
assign( vB, getVSReg( XB ) );
/* For all the VSX move instructions, the contents of doubleword element 1
* of VSX[XT] are undefined after the operation; therefore, we can simply
* move the entire array element where it makes sense to do so.
*/
switch (opc2) {
case 0x2B2: // xsabsdp (VSX scalar absolute value double-precision
{
/* Move abs val of dw 0 of VSX[XB] to dw 0 of VSX[XT]. */
IRTemp absVal = newTemp(Ity_V128);
assign(absVal, binop(Iop_ShrV128, binop(Iop_ShlV128, mkexpr(vB), mkU8(1)), mkU8(1)));
DIP("xsabsdp v%d,v%d\n", (UInt)XT, (UInt)XB);
putVSReg(XT, mkexpr(absVal));
break;
}
case 0x2C0: // xscpsgndp
{
/* Scalar copy sign double-precision */
IRTemp vecA_signbit = newTemp(Ity_V128);
IRTemp vecB_no_signbit = newTemp(Ity_V128);
IRTemp vec_result = newTemp(Ity_V128);
DIP("xscpsgndp v%d,v%d v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
assign( vecB_no_signbit, binop( Iop_ShrV128, binop( Iop_ShlV128,
mkexpr( vB ),
mkU8( 1 ) ),
mkU8( 1 ) ) );
assign( vecA_signbit, binop( Iop_ShlV128, binop( Iop_ShrV128,
mkexpr( vA ),
mkU8( 127 ) ),
mkU8( 127 ) ) );
assign( vec_result, binop( Iop_OrV128, mkexpr(vecA_signbit), mkexpr( vecB_no_signbit ) ) );
putVSReg(XT, mkexpr(vec_result));
break;
}
case 0x2D2: // xsnabsdp
{
/* Scalar negative absolute value double-precision */
IRTemp vec_neg_signbit = newTemp(Ity_V128);
DIP("xsnabsdp v%d,v%d\n", (UInt)XT, (UInt)XB);
assign( vec_neg_signbit, unop( Iop_NotV128, binop( Iop_ShrV128,
mkV128( 0xffff ),
mkU8( 1 ) ) ) );
putVSReg(XT, binop(Iop_OrV128, mkexpr(vec_neg_signbit), mkexpr(vB)));
break;
}
case 0x2F2: // xsnegdp
{
/* Scalar negate double-precision */
IRTemp vecB_no_signbit = newTemp(Ity_V128);
IRTemp vecB_signbit_comp = newTemp(Ity_V128);
DIP("xsnabsdp v%d,v%d\n", (UInt)XT, (UInt)XB);
assign( vecB_no_signbit, binop( Iop_ShrV128, binop( Iop_ShlV128,
mkexpr( vB ),
mkU8( 1 ) ),
mkU8( 1 ) ) );
assign( vecB_signbit_comp, binop( Iop_ShlV128,
unop( Iop_NotV128,
binop( Iop_ShrV128,
mkexpr( vB ),
mkU8( 127 ) ) ),
mkU8( 127 ) ) );
putVSReg( XT, binop( Iop_OrV128, mkexpr( vecB_no_signbit ),
mkexpr( vecB_signbit_comp ) ) );
break;
}
case 0x280: // xsmaxdp (VSX Scalar Maximum Double-Precision)
case 0x2A0: // xsmindp (VSX Scalar Minimum Double-Precision)
{
IRTemp frA = newTemp(Ity_I64);
IRTemp frB = newTemp(Ity_I64);
Bool isMin = opc2 == 0x2A0 ? True : False;
DIP("%s v%d,v%d v%d\n", isMin ? "xsmaxdp" : "xsmindp", (UInt)XT, (UInt)XA, (UInt)XB);
assign(frA, unop(Iop_V128HIto64, mkexpr( vA )));
assign(frB, unop(Iop_V128HIto64, mkexpr( vB )));
putVSReg( XT, binop( Iop_64HLtoV128, get_max_min_fp(frA, frB, isMin), mkU64( 0 ) ) );
break;
}
case 0x0F2: // xsrdpim (VSX Scalar Round to Double-Precision Integer using round toward -Infinity)
case 0x0D2: // xsrdpip (VSX Scalar Round to Double-Precision Integer using round toward +Infinity)
case 0x0D6: // xsrdpic (VSX Scalar Round to Double-Precision Integer using Current rounding mode)
case 0x0B2: // xsrdpiz (VSX Scalar Round to Double-Precision Integer using round toward Zero)
case 0x092: // xsrdpi (VSX Scalar Round to Double-Precision Integer using round toward Nearest Away)
{
IRTemp frB_I64 = newTemp(Ity_I64);
IRExpr * frD_fp_round = NULL;
UChar * insn_suffix = NULL;
assign(frB_I64, unop(Iop_V128HIto64, mkexpr( vB )));
frD_fp_round = _do_vsx_fp_roundToInt(frB_I64, opc2, insn_suffix);
DIP("xsrdpi%s v%d,v%d\n", insn_suffix, (UInt)XT, (UInt)XB);
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64, frD_fp_round),
mkU64( 0 ) ) );
break;
}
case 0x0B4: // xsredp (VSX Scalar Reciprocal Estimate Double-Precision)
case 0x094: // xsrsqrtedp (VSX Scalar Reciprocal Square Root Estimate Double-Precision)
{
IRTemp frB = newTemp(Ity_F64);
IRTemp sqrt = newTemp(Ity_F64);
IRExpr* ieee_one = IRExpr_Const(IRConst_F64i(0x3ff0000000000000ULL));
IRExpr* rm = get_IR_roundingmode();
Bool redp = opc2 == 0x0B4;
DIP("%s v%d,v%d\n", redp ? "xsredp" : "xsrsqrtedp", (UInt)XT, (UInt)XB);
assign( frB,
unop( Iop_ReinterpI64asF64,
unop( Iop_V128HIto64, mkexpr( vB ) ) ) );
if (!redp)
assign( sqrt,
binop( Iop_SqrtF64,
rm,
mkexpr(frB) ) );
putVSReg( XT,
binop( Iop_64HLtoV128,
unop( Iop_ReinterpF64asI64,
triop( Iop_DivF64,
rm,
ieee_one,
redp ? mkexpr( frB ) : mkexpr( sqrt ) ) ),
mkU64( 0 ) ) );
break;
}
default:
vex_printf( "dis_vxs_misc(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX Logical Instructions
*/
static Bool
dis_vx_logic ( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT ( theInstr );
UChar XA = ifieldRegXA ( theInstr );
UChar XB = ifieldRegXB ( theInstr );
IRTemp vA = newTemp( Ity_V128 );
IRTemp vB = newTemp( Ity_V128 );
if (opc1 != 0x3C) {
vex_printf( "dis_vx_logic(ppc)(instr)\n" );
return False;
}
assign( vA, getVSReg( XA ) );
assign( vB, getVSReg( XB ) );
switch (opc2) {
case 0x268: // xxlxor
DIP("xxlxor v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_XorV128, mkexpr( vA ), mkexpr( vB ) ) );
break;
case 0x248: // xxlor
DIP("xxlor v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_OrV128, mkexpr( vA ), mkexpr( vB ) ) );
break;
case 0x288: // xxlnor
DIP("xxlnor v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, unop( Iop_NotV128, binop( Iop_OrV128, mkexpr( vA ),
mkexpr( vB ) ) ) );
break;
case 0x208: // xxland
DIP("xxland v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_AndV128, mkexpr( vA ), mkexpr( vB ) ) );
break;
case 0x228: //xxlandc
DIP("xxlandc v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, binop( Iop_AndV128, mkexpr( vA ), unop( Iop_NotV128,
mkexpr( vB ) ) ) );
break;
default:
vex_printf( "dis_vx_logic(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX Load Instructions
* NOTE: VSX supports word-aligned storage access.
*/
static Bool
dis_vx_load ( UInt theInstr )
{
/* XX1-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT ( theInstr );
UChar rA_addr = ifieldRegA( theInstr );
UChar rB_addr = ifieldRegB( theInstr );
UInt opc2 = ifieldOPClo10( theInstr );
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp( ty );
if (opc1 != 0x1F) {
vex_printf( "dis_vx_load(ppc)(instr)\n" );
return False;
}
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
switch (opc2) {
case 0x24C: // lxsdx
{
IRExpr * exp;
DIP("lxsdx %d,r%u,r%u\n", (UInt)XT, rA_addr, rB_addr);
exp = loadBE( Ity_I64, mkexpr( EA ) );
// We need to pass an expression of type Ity_V128 with putVSReg, but the load
// we just performed is only a DW. But since the contents of VSR[XT] element 1
// are undefined after this operation, we can just do a splat op.
putVSReg( XT, binop( Iop_64HLtoV128, exp, exp ) );
break;
}
case 0x34C: // lxvd2x
{
IROp addOp = ty == Ity_I64 ? Iop_Add64 : Iop_Add32;
IRExpr * high, *low;
ULong ea_off = 8;
IRExpr* high_addr;
DIP("lxvd2x %d,r%u,r%u\n", (UInt)XT, rA_addr, rB_addr);
high = loadBE( Ity_I64, mkexpr( EA ) );
high_addr = binop( addOp, mkexpr( EA ), ty == Ity_I64 ? mkU64( ea_off )
: mkU32( ea_off ) );
low = loadBE( Ity_I64, high_addr );
putVSReg( XT, binop( Iop_64HLtoV128, high, low ) );
break;
}
case 0x14C: // lxvdsx
{
IRTemp data = newTemp(Ity_I64);
DIP("lxvdsx %d,r%u,r%u\n", (UInt)XT, rA_addr, rB_addr);
assign( data, loadBE( Ity_I64, mkexpr( EA ) ) );
putVSReg( XT, binop( Iop_64HLtoV128, mkexpr( data ), mkexpr( data ) ) );
break;
}
case 0x30C:
{
IRExpr * t3, *t2, *t1, *t0;
UInt ea_off = 0;
IRExpr* irx_addr;
DIP("lxvw4x %d,r%u,r%u\n", (UInt)XT, rA_addr, rB_addr);
t3 = loadBE( Ity_I32, mkexpr( EA ) );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
t2 = loadBE( Ity_I32, irx_addr );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
t1 = loadBE( Ity_I32, irx_addr );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
t0 = loadBE( Ity_I32, irx_addr );
putVSReg( XT, binop( Iop_64HLtoV128, binop( Iop_32HLto64, t3, t2 ),
binop( Iop_32HLto64, t1, t0 ) ) );
break;
}
default:
vex_printf( "dis_vx_load(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX Store Instructions
* NOTE: VSX supports word-aligned storage access.
*/
static Bool
dis_vx_store ( UInt theInstr )
{
/* XX1-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XS = ifieldRegXS( theInstr );
UChar rA_addr = ifieldRegA( theInstr );
UChar rB_addr = ifieldRegB( theInstr );
IRTemp vS = newTemp( Ity_V128 );
UInt opc2 = ifieldOPClo10( theInstr );
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp( ty );
if (opc1 != 0x1F) {
vex_printf( "dis_vx_store(ppc)(instr)\n" );
return False;
}
assign( EA, ea_rAor0_idxd( rA_addr, rB_addr ) );
assign( vS, getVSReg( XS ) );
switch (opc2) {
case 0x2CC:
{
IRExpr * high64;
DIP("stxsdx %d,r%u,r%u\n", (UInt)XS, rA_addr, rB_addr);
high64 = unop( Iop_V128HIto64, mkexpr( vS ) );
storeBE( mkexpr( EA ), high64 );
break;
}
case 0x3CC:
{
IRExpr * high64, *low64;
DIP("stxvd2x %d,r%u,r%u\n", (UInt)XS, rA_addr, rB_addr);
high64 = unop( Iop_V128HIto64, mkexpr( vS ) );
low64 = unop( Iop_V128to64, mkexpr( vS ) );
storeBE( mkexpr( EA ), high64 );
storeBE( binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ), ty == Ity_I64 ? mkU64( 8 )
: mkU32( 8 ) ), low64 );
break;
}
case 0x38C:
{
UInt ea_off = 0;
IRExpr* irx_addr;
IRTemp hi64 = newTemp( Ity_I64 );
IRTemp lo64 = newTemp( Ity_I64 );
DIP("stxvw4x %d,r%u,r%u\n", (UInt)XS, rA_addr, rB_addr);
// This instruction supports word-aligned stores, so EA may not be
// quad-word aligned. Therefore, do 4 individual word-size stores.
assign( hi64, unop( Iop_V128HIto64, mkexpr( vS ) ) );
assign( lo64, unop( Iop_V128to64, mkexpr( vS ) ) );
storeBE( mkexpr( EA ), unop( Iop_64HIto32, mkexpr( hi64 ) ) );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
storeBE( irx_addr, unop( Iop_64to32, mkexpr( hi64 ) ) );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
storeBE( irx_addr, unop( Iop_64HIto32, mkexpr( lo64 ) ) );
ea_off += 4;
irx_addr = binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ),
ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) );
storeBE( irx_addr, unop( Iop_64to32, mkexpr( lo64 ) ) );
break;
}
default:
vex_printf( "dis_vx_store(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
* VSX permute and other miscealleous instructions
*/
static Bool
dis_vx_permute_misc( UInt theInstr, UInt opc2 )
{
/* XX3-Form */
UChar opc1 = ifieldOPC( theInstr );
UChar XT = ifieldRegXT ( theInstr );
UChar XA = ifieldRegXA ( theInstr );
UChar XB = ifieldRegXB ( theInstr );
IRTemp vT = newTemp( Ity_V128 );
IRTemp vA = newTemp( Ity_V128 );
IRTemp vB = newTemp( Ity_V128 );
if (opc1 != 0x3C) {
vex_printf( "dis_vx_permute_misc(ppc)(instr)\n" );
return False;
}
assign( vA, getVSReg( XA ) );
assign( vB, getVSReg( XB ) );
switch (opc2) {
case 0x8: // xxsldwi (VSX Shift Left Double by Word Immediate)
{
UChar SHW = ifieldSHW ( theInstr );
IRTemp result = newTemp(Ity_V128);
if ( SHW != 0 ) {
IRTemp hi = newTemp(Ity_V128);
IRTemp lo = newTemp(Ity_V128);
assign( hi, binop(Iop_ShlV128, mkexpr(vA), mkU8(SHW*32)) );
assign( lo, binop(Iop_ShrV128, mkexpr(vB), mkU8(128-SHW*32)) );
assign ( result, binop(Iop_OrV128, mkexpr(hi), mkexpr(lo)) );
} else
assign ( result, mkexpr(vA) );
DIP("xxsldwi v%d,v%d,v%d,%d\n", (UInt)XT, (UInt)XA, (UInt)XB, (UInt)SHW);
putVSReg( XT, mkexpr(result) );
break;
}
case 0x28: // xpermdi (VSX Permute Doubleword Immediate)
{
UChar DM = ifieldDM ( theInstr );
IRTemp hi = newTemp(Ity_I64);
IRTemp lo = newTemp(Ity_I64);
if (DM & 0x2)
assign( hi, unop(Iop_V128to64, mkexpr(vA)) );
else
assign( hi, unop(Iop_V128HIto64, mkexpr(vA)) );
if (DM & 0x1)
assign( lo, unop(Iop_V128to64, mkexpr(vB)) );
else
assign( lo, unop(Iop_V128HIto64, mkexpr(vB)) );
assign( vT, binop(Iop_64HLtoV128, mkexpr(hi), mkexpr(lo)) );
DIP("xxpermdi v%d,v%d,v%d,0x%x\n", (UInt)XT, (UInt)XA, (UInt)XB, (UInt)DM);
putVSReg( XT, mkexpr( vT ) );
break;
}
case 0x48: // xxmrghw (VSX Merge High Word)
case 0xc8: // xxmrglw (VSX Merge Low Word)
{
char type = (opc2 == 0x48) ? 'h' : 'l';
IROp word_op = (opc2 == 0x48) ? Iop_V128HIto64 : Iop_V128to64;
IRTemp a64 = newTemp(Ity_I64);
IRTemp ahi32 = newTemp(Ity_I32);
IRTemp alo32 = newTemp(Ity_I32);
IRTemp b64 = newTemp(Ity_I64);
IRTemp bhi32 = newTemp(Ity_I32);
IRTemp blo32 = newTemp(Ity_I32);
assign( a64, unop(word_op, mkexpr(vA)) );
assign( ahi32, unop(Iop_64HIto32, mkexpr(a64)) );
assign( alo32, unop(Iop_64to32, mkexpr(a64)) );
assign( b64, unop(word_op, mkexpr(vB)) );
assign( bhi32, unop(Iop_64HIto32, mkexpr(b64)) );
assign( blo32, unop(Iop_64to32, mkexpr(b64)) );
assign( vT, binop(Iop_64HLtoV128,
binop(Iop_32HLto64, mkexpr(ahi32), mkexpr(bhi32)),
binop(Iop_32HLto64, mkexpr(alo32), mkexpr(blo32))) );
DIP("xxmrg%cw v%d,v%d,v%d\n", type, (UInt)XT, (UInt)XA, (UInt)XB);
putVSReg( XT, mkexpr( vT ) );
break;
}
case 0x018: // xxsel (VSX Select)
{
UChar XC = ifieldRegXC(theInstr);
IRTemp vC = newTemp( Ity_V128 );
assign( vC, getVSReg( XC ) );
DIP("xxsel v%d,v%d,v%d,v%d\n", (UInt)XT, (UInt)XA, (UInt)XB, (UInt)XC);
/* vD = (vA & ~vC) | (vB & vC) */
putVSReg( XT, binop(Iop_OrV128,
binop(Iop_AndV128, mkexpr(vA), unop(Iop_NotV128, mkexpr(vC))),
binop(Iop_AndV128, mkexpr(vB), mkexpr(vC))) );
break;
}
case 0x148: // xxspltw (VSX Splat Word)
{
UChar UIM = ifieldRegA(theInstr) & 3;
UChar sh_uim = (3 - (UIM)) * 32;
DIP("xxspltw v%d,v%d,%d\n", (UInt)XT, (UInt)XB, UIM);
putVSReg( XT,
unop( Iop_Dup32x4,
unop( Iop_V128to32,
binop( Iop_ShrV128, mkexpr( vB ), mkU8( sh_uim ) ) ) ) );
break;
}
default:
vex_printf( "dis_vx_permute_misc(ppc)(opc2)\n" );
return False;
}
return True;
}
/*
AltiVec Load Instructions
*/
static Bool dis_av_load ( VexAbiInfo* vbi, UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRTemp EA_align16 = newTemp(ty);
if (opc1 != 0x1F || b0 != 0) {
vex_printf("dis_av_load(ppc)(instr)\n");
return False;
}
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
assign( EA_align16, addr_align( mkexpr(EA), 16 ) );
switch (opc2) {
case 0x006: { // lvsl (Load Vector for Shift Left, AV p123)
IRDirty* d;
UInt vD_off = vectorGuestRegOffset(vD_addr);
IRExpr** args = mkIRExprVec_3(
mkU32(vD_off),
binop(Iop_And32, mkNarrowTo32(ty, mkexpr(EA)),
mkU32(0xF)),
mkU32(0)/*left*/ );
if (!mode64) {
d = unsafeIRDirty_0_N (
0/*regparms*/,
"ppc32g_dirtyhelper_LVS",
fnptr_to_fnentry(vbi, &ppc32g_dirtyhelper_LVS),
args );
} else {
d = unsafeIRDirty_0_N (
0/*regparms*/,
"ppc64g_dirtyhelper_LVS",
fnptr_to_fnentry(vbi, &ppc64g_dirtyhelper_LVS),
args );
}
DIP("lvsl v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
/* declare guest state effects */
d->needsBBP = True;
d->nFxState = 1;
vex_bzero(&d->fxState, sizeof(d->fxState));
d->fxState[0].fx = Ifx_Write;
d->fxState[0].offset = vD_off;
d->fxState[0].size = sizeof(U128);
/* execute the dirty call, side-effecting guest state */
stmt( IRStmt_Dirty(d) );
break;
}
case 0x026: { // lvsr (Load Vector for Shift Right, AV p125)
IRDirty* d;
UInt vD_off = vectorGuestRegOffset(vD_addr);
IRExpr** args = mkIRExprVec_3(
mkU32(vD_off),
binop(Iop_And32, mkNarrowTo32(ty, mkexpr(EA)),
mkU32(0xF)),
mkU32(1)/*right*/ );
if (!mode64) {
d = unsafeIRDirty_0_N (
0/*regparms*/,
"ppc32g_dirtyhelper_LVS",
fnptr_to_fnentry(vbi, &ppc32g_dirtyhelper_LVS),
args );
} else {
d = unsafeIRDirty_0_N (
0/*regparms*/,
"ppc64g_dirtyhelper_LVS",
fnptr_to_fnentry(vbi, &ppc64g_dirtyhelper_LVS),
args );
}
DIP("lvsr v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
/* declare guest state effects */
d->needsBBP = True;
d->nFxState = 1;
vex_bzero(&d->fxState, sizeof(d->fxState));
d->fxState[0].fx = Ifx_Write;
d->fxState[0].offset = vD_off;
d->fxState[0].size = sizeof(U128);
/* execute the dirty call, side-effecting guest state */
stmt( IRStmt_Dirty(d) );
break;
}
case 0x007: // lvebx (Load Vector Element Byte Indexed, AV p119)
DIP("lvebx v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
/* loads addressed byte into vector[EA[0:3]
since all other destination bytes are undefined,
can simply load entire vector from 16-aligned EA */
putVReg( vD_addr, loadBE(Ity_V128, mkexpr(EA_align16)) );
break;
case 0x027: // lvehx (Load Vector Element Half Word Indexed, AV p121)
DIP("lvehx v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
/* see note for lvebx */
putVReg( vD_addr, loadBE(Ity_V128, mkexpr(EA_align16)) );
break;
case 0x047: // lvewx (Load Vector Element Word Indexed, AV p122)
DIP("lvewx v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
/* see note for lvebx */
putVReg( vD_addr, loadBE(Ity_V128, mkexpr(EA_align16)) );
break;
case 0x067: // lvx (Load Vector Indexed, AV p127)
DIP("lvx v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
putVReg( vD_addr, loadBE(Ity_V128, mkexpr(EA_align16)) );
break;
case 0x167: // lvxl (Load Vector Indexed LRU, AV p128)
DIP("lvxl v%d,r%u,r%u\n", vD_addr, rA_addr, rB_addr);
putVReg( vD_addr, loadBE(Ity_V128, mkexpr(EA_align16)) );
break;
default:
vex_printf("dis_av_load(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Store Instructions
*/
static Bool dis_av_store ( UInt theInstr )
{
/* X-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vS_addr = ifieldRegDS(theInstr);
UChar rA_addr = ifieldRegA(theInstr);
UChar rB_addr = ifieldRegB(theInstr);
UInt opc2 = ifieldOPClo10(theInstr);
UChar b0 = ifieldBIT0(theInstr);
IRType ty = mode64 ? Ity_I64 : Ity_I32;
IRTemp EA = newTemp(ty);
IRTemp addr_aligned = newTemp(ty);
IRTemp vS = newTemp(Ity_V128);
IRTemp eb = newTemp(Ity_I8);
IRTemp idx = newTemp(Ity_I8);
if (opc1 != 0x1F || b0 != 0) {
vex_printf("dis_av_store(ppc)(instr)\n");
return False;
}
assign( vS, getVReg(vS_addr));
assign( EA, ea_rAor0_idxd(rA_addr, rB_addr) );
switch (opc2) {
case 0x087: { // stvebx (Store Vector Byte Indexed, AV p131)
DIP("stvebx v%d,r%u,r%u\n", vS_addr, rA_addr, rB_addr);
assign( eb, binop(Iop_And8, mkU8(0xF),
unop(Iop_32to8,
mkNarrowTo32(ty, mkexpr(EA)) )) );
assign( idx, binop(Iop_Shl8,
binop(Iop_Sub8, mkU8(15), mkexpr(eb)),
mkU8(3)) );
storeBE( mkexpr(EA),
unop(Iop_32to8, unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vS), mkexpr(idx)))) );
break;
}
case 0x0A7: { // stvehx (Store Vector Half Word Indexed, AV p132)
DIP("stvehx v%d,r%u,r%u\n", vS_addr, rA_addr, rB_addr);
assign( addr_aligned, addr_align(mkexpr(EA), 2) );
assign( eb, binop(Iop_And8, mkU8(0xF),
mkNarrowTo8(ty, mkexpr(addr_aligned) )) );
assign( idx, binop(Iop_Shl8,
binop(Iop_Sub8, mkU8(14), mkexpr(eb)),
mkU8(3)) );
storeBE( mkexpr(addr_aligned),
unop(Iop_32to16, unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vS), mkexpr(idx)))) );
break;
}
case 0x0C7: { // stvewx (Store Vector Word Indexed, AV p133)
DIP("stvewx v%d,r%u,r%u\n", vS_addr, rA_addr, rB_addr);
assign( addr_aligned, addr_align(mkexpr(EA), 4) );
assign( eb, binop(Iop_And8, mkU8(0xF),
mkNarrowTo8(ty, mkexpr(addr_aligned) )) );
assign( idx, binop(Iop_Shl8,
binop(Iop_Sub8, mkU8(12), mkexpr(eb)),
mkU8(3)) );
storeBE( mkexpr(addr_aligned),
unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vS), mkexpr(idx))) );
break;
}
case 0x0E7: // stvx (Store Vector Indexed, AV p134)
DIP("stvx v%d,r%u,r%u\n", vS_addr, rA_addr, rB_addr);
storeBE( addr_align( mkexpr(EA), 16 ), mkexpr(vS) );
break;
case 0x1E7: // stvxl (Store Vector Indexed LRU, AV p135)
DIP("stvxl v%d,r%u,r%u\n", vS_addr, rA_addr, rB_addr);
storeBE( addr_align( mkexpr(EA), 16 ), mkexpr(vS) );
break;
default:
vex_printf("dis_av_store(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Arithmetic Instructions
*/
static Bool dis_av_arith ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp z3 = newTemp(Ity_I64);
IRTemp z2 = newTemp(Ity_I64);
IRTemp z1 = newTemp(Ity_I64);
IRTemp z0 = newTemp(Ity_I64);
IRTemp aEvn, aOdd;
IRTemp a15, a14, a13, a12, a11, a10, a9, a8;
IRTemp a7, a6, a5, a4, a3, a2, a1, a0;
IRTemp b3, b2, b1, b0;
aEvn = aOdd = IRTemp_INVALID;
a15 = a14 = a13 = a12 = a11 = a10 = a9 = a8 = IRTemp_INVALID;
a7 = a6 = a5 = a4 = a3 = a2 = a1 = a0 = IRTemp_INVALID;
b3 = b2 = b1 = b0 = IRTemp_INVALID;
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_arith(ppc)(opc1 != 0x4)\n");
return False;
}
switch (opc2) {
/* Add */
case 0x180: { // vaddcuw (Add Carryout Unsigned Word, AV p136)
DIP("vaddcuw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* unsigned_ov(x+y) = (y >u not(x)) */
putVReg( vD_addr, binop(Iop_ShrN32x4,
binop(Iop_CmpGT32Ux4, mkexpr(vB),
unop(Iop_NotV128, mkexpr(vA))),
mkU8(31)) );
break;
}
case 0x000: // vaddubm (Add Unsigned Byte Modulo, AV p141)
DIP("vaddubm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Add8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x040: // vadduhm (Add Unsigned Half Word Modulo, AV p143)
DIP("vadduhm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Add16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x080: // vadduwm (Add Unsigned Word Modulo, AV p145)
DIP("vadduwm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Add32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x200: // vaddubs (Add Unsigned Byte Saturate, AV p142)
DIP("vaddubs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd8Ux16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT], perhaps via new primop: Iop_SatOfQAdd8Ux16
break;
case 0x240: // vadduhs (Add Unsigned Half Word Saturate, AV p144)
DIP("vadduhs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd16Ux8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x280: // vadduws (Add Unsigned Word Saturate, AV p146)
DIP("vadduws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd32Ux4, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x300: // vaddsbs (Add Signed Byte Saturate, AV p138)
DIP("vaddsbs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd8Sx16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x340: // vaddshs (Add Signed Half Word Saturate, AV p139)
DIP("vaddshs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd16Sx8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x380: // vaddsws (Add Signed Word Saturate, AV p140)
DIP("vaddsws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QAdd32Sx4, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
/* Subtract */
case 0x580: { // vsubcuw (Subtract Carryout Unsigned Word, AV p260)
DIP("vsubcuw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* unsigned_ov(x-y) = (y >u x) */
putVReg( vD_addr, binop(Iop_ShrN32x4,
unop(Iop_NotV128,
binop(Iop_CmpGT32Ux4, mkexpr(vB),
mkexpr(vA))),
mkU8(31)) );
break;
}
case 0x400: // vsububm (Subtract Unsigned Byte Modulo, AV p265)
DIP("vsububm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sub8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x440: // vsubuhm (Subtract Unsigned Half Word Modulo, AV p267)
DIP("vsubuhm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sub16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x480: // vsubuwm (Subtract Unsigned Word Modulo, AV p269)
DIP("vsubuwm v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sub32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x600: // vsububs (Subtract Unsigned Byte Saturate, AV p266)
DIP("vsububs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub8Ux16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x640: // vsubuhs (Subtract Unsigned HWord Saturate, AV p268)
DIP("vsubuhs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub16Ux8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x680: // vsubuws (Subtract Unsigned Word Saturate, AV p270)
DIP("vsubuws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub32Ux4, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x700: // vsubsbs (Subtract Signed Byte Saturate, AV p262)
DIP("vsubsbs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub8Sx16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x740: // vsubshs (Subtract Signed Half Word Saturate, AV p263)
DIP("vsubshs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub16Sx8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
case 0x780: // vsubsws (Subtract Signed Word Saturate, AV p264)
DIP("vsubsws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_QSub32Sx4, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
break;
/* Maximum */
case 0x002: // vmaxub (Maximum Unsigned Byte, AV p182)
DIP("vmaxub v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max8Ux16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x042: // vmaxuh (Maximum Unsigned Half Word, AV p183)
DIP("vmaxuh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max16Ux8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x082: // vmaxuw (Maximum Unsigned Word, AV p184)
DIP("vmaxuw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max32Ux4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x102: // vmaxsb (Maximum Signed Byte, AV p179)
DIP("vmaxsb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max8Sx16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x142: // vmaxsh (Maximum Signed Half Word, AV p180)
DIP("vmaxsh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max16Sx8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x182: // vmaxsw (Maximum Signed Word, AV p181)
DIP("vmaxsw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max32Sx4, mkexpr(vA), mkexpr(vB)) );
break;
/* Minimum */
case 0x202: // vminub (Minimum Unsigned Byte, AV p191)
DIP("vminub v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min8Ux16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x242: // vminuh (Minimum Unsigned Half Word, AV p192)
DIP("vminuh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min16Ux8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x282: // vminuw (Minimum Unsigned Word, AV p193)
DIP("vminuw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min32Ux4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x302: // vminsb (Minimum Signed Byte, AV p188)
DIP("vminsb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min8Sx16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x342: // vminsh (Minimum Signed Half Word, AV p189)
DIP("vminsh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min16Sx8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x382: // vminsw (Minimum Signed Word, AV p190)
DIP("vminsw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min32Sx4, mkexpr(vA), mkexpr(vB)) );
break;
/* Average */
case 0x402: // vavgub (Average Unsigned Byte, AV p152)
DIP("vavgub v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg8Ux16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x442: // vavguh (Average Unsigned Half Word, AV p153)
DIP("vavguh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg16Ux8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x482: // vavguw (Average Unsigned Word, AV p154)
DIP("vavguw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg32Ux4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x502: // vavgsb (Average Signed Byte, AV p149)
DIP("vavgsb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg8Sx16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x542: // vavgsh (Average Signed Half Word, AV p150)
DIP("vavgsh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg16Sx8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x582: // vavgsw (Average Signed Word, AV p151)
DIP("vavgsw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Avg32Sx4, mkexpr(vA), mkexpr(vB)) );
break;
/* Multiply */
case 0x008: // vmuloub (Multiply Odd Unsigned Byte, AV p213)
DIP("vmuloub v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_MullEven8Ux16, mkexpr(vA), mkexpr(vB)));
break;
case 0x048: // vmulouh (Multiply Odd Unsigned Half Word, AV p214)
DIP("vmulouh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_MullEven16Ux8, mkexpr(vA), mkexpr(vB)));
break;
case 0x108: // vmulosb (Multiply Odd Signed Byte, AV p211)
DIP("vmulosb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_MullEven8Sx16, mkexpr(vA), mkexpr(vB)));
break;
case 0x148: // vmulosh (Multiply Odd Signed Half Word, AV p212)
DIP("vmulosh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_MullEven16Sx8, mkexpr(vA), mkexpr(vB)));
break;
case 0x208: // vmuleub (Multiply Even Unsigned Byte, AV p209)
DIP("vmuleub v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, MK_Iop_MullOdd8Ux16( mkexpr(vA), mkexpr(vB) ));
break;
case 0x248: // vmuleuh (Multiply Even Unsigned Half Word, AV p210)
DIP("vmuleuh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, MK_Iop_MullOdd16Ux8( mkexpr(vA), mkexpr(vB) ));
break;
case 0x308: // vmulesb (Multiply Even Signed Byte, AV p207)
DIP("vmulesb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, MK_Iop_MullOdd8Sx16( mkexpr(vA), mkexpr(vB) ));
break;
case 0x348: // vmulesh (Multiply Even Signed Half Word, AV p208)
DIP("vmulesh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, MK_Iop_MullOdd16Sx8( mkexpr(vA), mkexpr(vB) ));
break;
/* Sum Across Partial */
case 0x608: { // vsum4ubs (Sum Partial (1/4) UB Saturate, AV p275)
IRTemp aEE, aEO, aOE, aOO;
aEE = aEO = aOE = aOO = IRTemp_INVALID;
DIP("vsum4ubs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* vA: V128_8Ux16 -> 4 x V128_32Ux4, sign-extended */
expand8Ux16( mkexpr(vA), &aEvn, &aOdd ); // (15,13...),(14,12...)
expand16Ux8( mkexpr(aEvn), &aEE, &aEO ); // (15,11...),(13, 9...)
expand16Ux8( mkexpr(aOdd), &aOE, &aOO ); // (14,10...),(12, 8...)
/* break V128 to 4xI32's, zero-extending to I64's */
breakV128to4x64U( mkexpr(aEE), &a15, &a11, &a7, &a3 );
breakV128to4x64U( mkexpr(aOE), &a14, &a10, &a6, &a2 );
breakV128to4x64U( mkexpr(aEO), &a13, &a9, &a5, &a1 );
breakV128to4x64U( mkexpr(aOO), &a12, &a8, &a4, &a0 );
breakV128to4x64U( mkexpr(vB), &b3, &b2, &b1, &b0 );
/* add lanes */
assign( z3, binop(Iop_Add64, mkexpr(b3),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a15), mkexpr(a14)),
binop(Iop_Add64, mkexpr(a13), mkexpr(a12)))) );
assign( z2, binop(Iop_Add64, mkexpr(b2),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a11), mkexpr(a10)),
binop(Iop_Add64, mkexpr(a9), mkexpr(a8)))) );
assign( z1, binop(Iop_Add64, mkexpr(b1),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a7), mkexpr(a6)),
binop(Iop_Add64, mkexpr(a5), mkexpr(a4)))) );
assign( z0, binop(Iop_Add64, mkexpr(b0),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a3), mkexpr(a2)),
binop(Iop_Add64, mkexpr(a1), mkexpr(a0)))) );
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64U( mkexpr(z3), mkexpr(z2),
mkexpr(z1), mkexpr(z0)) );
break;
}
case 0x708: { // vsum4sbs (Sum Partial (1/4) SB Saturate, AV p273)
IRTemp aEE, aEO, aOE, aOO;
aEE = aEO = aOE = aOO = IRTemp_INVALID;
DIP("vsum4sbs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* vA: V128_8Sx16 -> 4 x V128_32Sx4, sign-extended */
expand8Sx16( mkexpr(vA), &aEvn, &aOdd ); // (15,13...),(14,12...)
expand16Sx8( mkexpr(aEvn), &aEE, &aEO ); // (15,11...),(13, 9...)
expand16Sx8( mkexpr(aOdd), &aOE, &aOO ); // (14,10...),(12, 8...)
/* break V128 to 4xI32's, sign-extending to I64's */
breakV128to4x64S( mkexpr(aEE), &a15, &a11, &a7, &a3 );
breakV128to4x64S( mkexpr(aOE), &a14, &a10, &a6, &a2 );
breakV128to4x64S( mkexpr(aEO), &a13, &a9, &a5, &a1 );
breakV128to4x64S( mkexpr(aOO), &a12, &a8, &a4, &a0 );
breakV128to4x64S( mkexpr(vB), &b3, &b2, &b1, &b0 );
/* add lanes */
assign( z3, binop(Iop_Add64, mkexpr(b3),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a15), mkexpr(a14)),
binop(Iop_Add64, mkexpr(a13), mkexpr(a12)))) );
assign( z2, binop(Iop_Add64, mkexpr(b2),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a11), mkexpr(a10)),
binop(Iop_Add64, mkexpr(a9), mkexpr(a8)))) );
assign( z1, binop(Iop_Add64, mkexpr(b1),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a7), mkexpr(a6)),
binop(Iop_Add64, mkexpr(a5), mkexpr(a4)))) );
assign( z0, binop(Iop_Add64, mkexpr(b0),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a3), mkexpr(a2)),
binop(Iop_Add64, mkexpr(a1), mkexpr(a0)))) );
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64S( mkexpr(z3), mkexpr(z2),
mkexpr(z1), mkexpr(z0)) );
break;
}
case 0x648: { // vsum4shs (Sum Partial (1/4) SHW Saturate, AV p274)
DIP("vsum4shs v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* vA: V128_16Sx8 -> 2 x V128_32Sx4, sign-extended */
expand16Sx8( mkexpr(vA), &aEvn, &aOdd ); // (7,5...),(6,4...)
/* break V128 to 4xI32's, sign-extending to I64's */
breakV128to4x64S( mkexpr(aEvn), &a7, &a5, &a3, &a1 );
breakV128to4x64S( mkexpr(aOdd), &a6, &a4, &a2, &a0 );
breakV128to4x64S( mkexpr(vB), &b3, &b2, &b1, &b0 );
/* add lanes */
assign( z3, binop(Iop_Add64, mkexpr(b3),
binop(Iop_Add64, mkexpr(a7), mkexpr(a6))));
assign( z2, binop(Iop_Add64, mkexpr(b2),
binop(Iop_Add64, mkexpr(a5), mkexpr(a4))));
assign( z1, binop(Iop_Add64, mkexpr(b1),
binop(Iop_Add64, mkexpr(a3), mkexpr(a2))));
assign( z0, binop(Iop_Add64, mkexpr(b0),
binop(Iop_Add64, mkexpr(a1), mkexpr(a0))));
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64S( mkexpr(z3), mkexpr(z2),
mkexpr(z1), mkexpr(z0)) );
break;
}
case 0x688: { // vsum2sws (Sum Partial (1/2) SW Saturate, AV p272)
DIP("vsum2sws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* break V128 to 4xI32's, sign-extending to I64's */
breakV128to4x64S( mkexpr(vA), &a3, &a2, &a1, &a0 );
breakV128to4x64S( mkexpr(vB), &b3, &b2, &b1, &b0 );
/* add lanes */
assign( z2, binop(Iop_Add64, mkexpr(b2),
binop(Iop_Add64, mkexpr(a3), mkexpr(a2))) );
assign( z0, binop(Iop_Add64, mkexpr(b0),
binop(Iop_Add64, mkexpr(a1), mkexpr(a0))) );
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64S( mkU64(0), mkexpr(z2),
mkU64(0), mkexpr(z0)) );
break;
}
case 0x788: { // vsumsws (Sum SW Saturate, AV p271)
DIP("vsumsws v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
/* break V128 to 4xI32's, sign-extending to I64's */
breakV128to4x64S( mkexpr(vA), &a3, &a2, &a1, &a0 );
breakV128to4x64S( mkexpr(vB), &b3, &b2, &b1, &b0 );
/* add lanes */
assign( z0, binop(Iop_Add64, mkexpr(b0),
binop(Iop_Add64,
binop(Iop_Add64, mkexpr(a3), mkexpr(a2)),
binop(Iop_Add64, mkexpr(a1), mkexpr(a0)))) );
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64S( mkU64(0), mkU64(0),
mkU64(0), mkexpr(z0)) );
break;
}
default:
vex_printf("dis_av_arith(ppc)(opc2=0x%x)\n", opc2);
return False;
}
return True;
}
/*
AltiVec Logic Instructions
*/
static Bool dis_av_logic ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_logic(ppc)(opc1 != 0x4)\n");
return False;
}
switch (opc2) {
case 0x404: // vand (And, AV p147)
DIP("vand v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_AndV128, mkexpr(vA), mkexpr(vB)) );
break;
case 0x444: // vandc (And, AV p148)
DIP("vandc v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_AndV128, mkexpr(vA),
unop(Iop_NotV128, mkexpr(vB))) );
break;
case 0x484: // vor (Or, AV p217)
DIP("vor v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_OrV128, mkexpr(vA), mkexpr(vB)) );
break;
case 0x4C4: // vxor (Xor, AV p282)
DIP("vxor v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_XorV128, mkexpr(vA), mkexpr(vB)) );
break;
case 0x504: // vnor (Nor, AV p216)
DIP("vnor v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
unop(Iop_NotV128, binop(Iop_OrV128, mkexpr(vA), mkexpr(vB))) );
break;
default:
vex_printf("dis_av_logic(ppc)(opc2=0x%x)\n", opc2);
return False;
}
return True;
}
/*
AltiVec Compare Instructions
*/
static Bool dis_av_cmp ( UInt theInstr )
{
/* VXR-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UChar flag_rC = ifieldBIT10(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 10 );
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp vD = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_cmp(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x006: // vcmpequb (Compare Equal-to Unsigned B, AV p160)
DIP("vcmpequb%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpEQ8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x046: // vcmpequh (Compare Equal-to Unsigned HW, AV p161)
DIP("vcmpequh%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpEQ16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x086: // vcmpequw (Compare Equal-to Unsigned W, AV p162)
DIP("vcmpequw%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpEQ32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x206: // vcmpgtub (Compare Greater-than Unsigned B, AV p168)
DIP("vcmpgtub%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT8Ux16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x246: // vcmpgtuh (Compare Greater-than Unsigned HW, AV p169)
DIP("vcmpgtuh%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT16Ux8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x286: // vcmpgtuw (Compare Greater-than Unsigned W, AV p170)
DIP("vcmpgtuw%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT32Ux4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x306: // vcmpgtsb (Compare Greater-than Signed B, AV p165)
DIP("vcmpgtsb%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT8Sx16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x346: // vcmpgtsh (Compare Greater-than Signed HW, AV p166)
DIP("vcmpgtsh%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT16Sx8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x386: // vcmpgtsw (Compare Greater-than Signed W, AV p167)
DIP("vcmpgtsw%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT32Sx4, mkexpr(vA), mkexpr(vB)) );
break;
default:
vex_printf("dis_av_cmp(ppc)(opc2)\n");
return False;
}
putVReg( vD_addr, mkexpr(vD) );
if (flag_rC) {
set_AV_CR6( mkexpr(vD), True );
}
return True;
}
/*
AltiVec Multiply-Sum Instructions
*/
static Bool dis_av_multarith ( UInt theInstr )
{
/* VA-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UChar vC_addr = ifieldRegC(theInstr);
UChar opc2 = toUChar( IFIELD( theInstr, 0, 6 ) );
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp vC = newTemp(Ity_V128);
IRTemp zeros = newTemp(Ity_V128);
IRTemp aLo = newTemp(Ity_V128);
IRTemp bLo = newTemp(Ity_V128);
IRTemp cLo = newTemp(Ity_V128);
IRTemp zLo = newTemp(Ity_V128);
IRTemp aHi = newTemp(Ity_V128);
IRTemp bHi = newTemp(Ity_V128);
IRTemp cHi = newTemp(Ity_V128);
IRTemp zHi = newTemp(Ity_V128);
IRTemp abEvn = newTemp(Ity_V128);
IRTemp abOdd = newTemp(Ity_V128);
IRTemp z3 = newTemp(Ity_I64);
IRTemp z2 = newTemp(Ity_I64);
IRTemp z1 = newTemp(Ity_I64);
IRTemp z0 = newTemp(Ity_I64);
IRTemp ab7, ab6, ab5, ab4, ab3, ab2, ab1, ab0;
IRTemp c3, c2, c1, c0;
ab7 = ab6 = ab5 = ab4 = ab3 = ab2 = ab1 = ab0 = IRTemp_INVALID;
c3 = c2 = c1 = c0 = IRTemp_INVALID;
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
assign( vC, getVReg(vC_addr));
assign( zeros, unop(Iop_Dup32x4, mkU32(0)) );
if (opc1 != 0x4) {
vex_printf("dis_av_multarith(ppc)(instr)\n");
return False;
}
switch (opc2) {
/* Multiply-Add */
case 0x20: { // vmhaddshs (Mult Hi, Add Signed HW Saturate, AV p185)
IRTemp cSigns = newTemp(Ity_V128);
DIP("vmhaddshs v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
assign(cSigns, binop(Iop_CmpGT16Sx8, mkexpr(zeros), mkexpr(vC)));
assign(aLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vA)));
assign(bLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vB)));
assign(cLo, binop(Iop_InterleaveLO16x8, mkexpr(cSigns),mkexpr(vC)));
assign(aHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vA)));
assign(bHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vB)));
assign(cHi, binop(Iop_InterleaveHI16x8, mkexpr(cSigns),mkexpr(vC)));
assign( zLo, binop(Iop_Add32x4, mkexpr(cLo),
binop(Iop_SarN32x4,
binop(Iop_MullEven16Sx8,
mkexpr(aLo), mkexpr(bLo)),
mkU8(15))) );
assign( zHi, binop(Iop_Add32x4, mkexpr(cHi),
binop(Iop_SarN32x4,
binop(Iop_MullEven16Sx8,
mkexpr(aHi), mkexpr(bHi)),
mkU8(15))) );
putVReg( vD_addr,
binop(Iop_QNarrowBin32Sto16Sx8, mkexpr(zHi), mkexpr(zLo)) );
break;
}
case 0x21: { // vmhraddshs (Mult High Round, Add Signed HW Saturate, AV p186)
IRTemp zKonst = newTemp(Ity_V128);
IRTemp cSigns = newTemp(Ity_V128);
DIP("vmhraddshs v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
assign(cSigns, binop(Iop_CmpGT16Sx8, mkexpr(zeros), mkexpr(vC)) );
assign(aLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vA)));
assign(bLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vB)));
assign(cLo, binop(Iop_InterleaveLO16x8, mkexpr(cSigns),mkexpr(vC)));
assign(aHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vA)));
assign(bHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vB)));
assign(cHi, binop(Iop_InterleaveHI16x8, mkexpr(cSigns),mkexpr(vC)));
/* shifting our const avoids store/load version of Dup */
assign( zKonst, binop(Iop_ShlN32x4, unop(Iop_Dup32x4, mkU32(0x1)),
mkU8(14)) );
assign( zLo, binop(Iop_Add32x4, mkexpr(cLo),
binop(Iop_SarN32x4,
binop(Iop_Add32x4, mkexpr(zKonst),
binop(Iop_MullEven16Sx8,
mkexpr(aLo), mkexpr(bLo))),
mkU8(15))) );
assign( zHi, binop(Iop_Add32x4, mkexpr(cHi),
binop(Iop_SarN32x4,
binop(Iop_Add32x4, mkexpr(zKonst),
binop(Iop_MullEven16Sx8,
mkexpr(aHi), mkexpr(bHi))),
mkU8(15))) );
putVReg( vD_addr,
binop(Iop_QNarrowBin32Sto16Sx8, mkexpr(zHi), mkexpr(zLo)) );
break;
}
case 0x22: { // vmladduhm (Mult Low, Add Unsigned HW Modulo, AV p194)
DIP("vmladduhm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
assign(aLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vA)));
assign(bLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vB)));
assign(cLo, binop(Iop_InterleaveLO16x8, mkexpr(zeros), mkexpr(vC)));
assign(aHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vA)));
assign(bHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vB)));
assign(cHi, binop(Iop_InterleaveHI16x8, mkexpr(zeros), mkexpr(vC)));
assign(zLo, binop(Iop_Add32x4,
binop(Iop_MullEven16Ux8, mkexpr(aLo), mkexpr(bLo)),
mkexpr(cLo)) );
assign(zHi, binop(Iop_Add32x4,
binop(Iop_MullEven16Ux8, mkexpr(aHi), mkexpr(bHi)),
mkexpr(cHi)));
putVReg( vD_addr,
binop(Iop_NarrowBin32to16x8, mkexpr(zHi), mkexpr(zLo)) );
break;
}
/* Multiply-Sum */
case 0x24: { // vmsumubm (Multiply Sum Unsigned B Modulo, AV p204)
IRTemp abEE, abEO, abOE, abOO;
abEE = abEO = abOE = abOO = IRTemp_INVALID;
DIP("vmsumubm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
/* multiply vA,vB (unsigned, widening) */
assign( abEvn, MK_Iop_MullOdd8Ux16( mkexpr(vA), mkexpr(vB) ));
assign( abOdd, binop(Iop_MullEven8Ux16, mkexpr(vA), mkexpr(vB)) );
/* evn,odd: V128_16Ux8 -> 2 x V128_32Ux4, zero-extended */
expand16Ux8( mkexpr(abEvn), &abEE, &abEO );
expand16Ux8( mkexpr(abOdd), &abOE, &abOO );
putVReg( vD_addr,
binop(Iop_Add32x4, mkexpr(vC),
binop(Iop_Add32x4,
binop(Iop_Add32x4, mkexpr(abEE), mkexpr(abEO)),
binop(Iop_Add32x4, mkexpr(abOE), mkexpr(abOO)))) );
break;
}
case 0x25: { // vmsummbm (Multiply Sum Mixed-Sign B Modulo, AV p201)
IRTemp aEvn, aOdd, bEvn, bOdd;
IRTemp abEE = newTemp(Ity_V128);
IRTemp abEO = newTemp(Ity_V128);
IRTemp abOE = newTemp(Ity_V128);
IRTemp abOO = newTemp(Ity_V128);
aEvn = aOdd = bEvn = bOdd = IRTemp_INVALID;
DIP("vmsummbm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
/* sign-extend vA, zero-extend vB, for mixed-sign multiply
(separating out adjacent lanes to different vectors) */
expand8Sx16( mkexpr(vA), &aEvn, &aOdd );
expand8Ux16( mkexpr(vB), &bEvn, &bOdd );
/* multiply vA, vB, again separating adjacent lanes */
assign( abEE, MK_Iop_MullOdd16Sx8( mkexpr(aEvn), mkexpr(bEvn) ));
assign( abEO, binop(Iop_MullEven16Sx8, mkexpr(aEvn), mkexpr(bEvn)) );
assign( abOE, MK_Iop_MullOdd16Sx8( mkexpr(aOdd), mkexpr(bOdd) ));
assign( abOO, binop(Iop_MullEven16Sx8, mkexpr(aOdd), mkexpr(bOdd)) );
/* add results together, + vC */
putVReg( vD_addr,
binop(Iop_QAdd32Sx4, mkexpr(vC),
binop(Iop_QAdd32Sx4,
binop(Iop_QAdd32Sx4, mkexpr(abEE), mkexpr(abEO)),
binop(Iop_QAdd32Sx4, mkexpr(abOE), mkexpr(abOO)))) );
break;
}
case 0x26: { // vmsumuhm (Multiply Sum Unsigned HW Modulo, AV p205)
DIP("vmsumuhm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
assign( abEvn, MK_Iop_MullOdd16Ux8( mkexpr(vA), mkexpr(vB) ));
assign( abOdd, binop(Iop_MullEven16Ux8, mkexpr(vA), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_Add32x4, mkexpr(vC),
binop(Iop_Add32x4, mkexpr(abEvn), mkexpr(abOdd))) );
break;
}
case 0x27: { // vmsumuhs (Multiply Sum Unsigned HW Saturate, AV p206)
DIP("vmsumuhs v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
/* widening multiply, separating lanes */
assign( abEvn, MK_Iop_MullOdd16Ux8(mkexpr(vA), mkexpr(vB) ));
assign( abOdd, binop(Iop_MullEven16Ux8, mkexpr(vA), mkexpr(vB)) );
/* break V128 to 4xI32's, zero-extending to I64's */
breakV128to4x64U( mkexpr(abEvn), &ab7, &ab5, &ab3, &ab1 );
breakV128to4x64U( mkexpr(abOdd), &ab6, &ab4, &ab2, &ab0 );
breakV128to4x64U( mkexpr(vC), &c3, &c2, &c1, &c0 );
/* add lanes */
assign( z3, binop(Iop_Add64, mkexpr(c3),
binop(Iop_Add64, mkexpr(ab7), mkexpr(ab6))));
assign( z2, binop(Iop_Add64, mkexpr(c2),
binop(Iop_Add64, mkexpr(ab5), mkexpr(ab4))));
assign( z1, binop(Iop_Add64, mkexpr(c1),
binop(Iop_Add64, mkexpr(ab3), mkexpr(ab2))));
assign( z0, binop(Iop_Add64, mkexpr(c0),
binop(Iop_Add64, mkexpr(ab1), mkexpr(ab0))));
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64U( mkexpr(z3), mkexpr(z2),
mkexpr(z1), mkexpr(z0)) );
break;
}
case 0x28: { // vmsumshm (Multiply Sum Signed HW Modulo, AV p202)
DIP("vmsumshm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
assign( abEvn, MK_Iop_MullOdd16Sx8( mkexpr(vA), mkexpr(vB) ));
assign( abOdd, binop(Iop_MullEven16Sx8, mkexpr(vA), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_Add32x4, mkexpr(vC),
binop(Iop_Add32x4, mkexpr(abOdd), mkexpr(abEvn))) );
break;
}
case 0x29: { // vmsumshs (Multiply Sum Signed HW Saturate, AV p203)
DIP("vmsumshs v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
/* widening multiply, separating lanes */
assign( abEvn, MK_Iop_MullOdd16Sx8( mkexpr(vA), mkexpr(vB) ));
assign( abOdd, binop(Iop_MullEven16Sx8, mkexpr(vA), mkexpr(vB)) );
/* break V128 to 4xI32's, sign-extending to I64's */
breakV128to4x64S( mkexpr(abEvn), &ab7, &ab5, &ab3, &ab1 );
breakV128to4x64S( mkexpr(abOdd), &ab6, &ab4, &ab2, &ab0 );
breakV128to4x64S( mkexpr(vC), &c3, &c2, &c1, &c0 );
/* add lanes */
assign( z3, binop(Iop_Add64, mkexpr(c3),
binop(Iop_Add64, mkexpr(ab7), mkexpr(ab6))));
assign( z2, binop(Iop_Add64, mkexpr(c2),
binop(Iop_Add64, mkexpr(ab5), mkexpr(ab4))));
assign( z1, binop(Iop_Add64, mkexpr(c1),
binop(Iop_Add64, mkexpr(ab3), mkexpr(ab2))));
assign( z0, binop(Iop_Add64, mkexpr(c0),
binop(Iop_Add64, mkexpr(ab1), mkexpr(ab0))));
/* saturate-narrow to 32bit, and combine to V128 */
putVReg( vD_addr, mkV128from4x64S( mkexpr(z3), mkexpr(z2),
mkexpr(z1), mkexpr(z0)) );
break;
}
default:
vex_printf("dis_av_multarith(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Shift/Rotate Instructions
*/
static Bool dis_av_shift ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4){
vex_printf("dis_av_shift(ppc)(instr)\n");
return False;
}
switch (opc2) {
/* Rotate */
case 0x004: // vrlb (Rotate Left Integer B, AV p234)
DIP("vrlb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Rol8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x044: // vrlh (Rotate Left Integer HW, AV p235)
DIP("vrlh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Rol16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x084: // vrlw (Rotate Left Integer W, AV p236)
DIP("vrlw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Rol32x4, mkexpr(vA), mkexpr(vB)) );
break;
/* Shift Left */
case 0x104: // vslb (Shift Left Integer B, AV p240)
DIP("vslb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shl8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x144: // vslh (Shift Left Integer HW, AV p242)
DIP("vslh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shl16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x184: // vslw (Shift Left Integer W, AV p244)
DIP("vslw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shl32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x1C4: { // vsl (Shift Left, AV p239)
IRTemp sh = newTemp(Ity_I8);
DIP("vsl v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( sh, binop(Iop_And8, mkU8(0x7),
unop(Iop_32to8,
unop(Iop_V128to32, mkexpr(vB)))) );
putVReg( vD_addr,
binop(Iop_ShlV128, mkexpr(vA), mkexpr(sh)) );
break;
}
case 0x40C: { // vslo (Shift Left by Octet, AV p243)
IRTemp sh = newTemp(Ity_I8);
DIP("vslo v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( sh, binop(Iop_And8, mkU8(0x78),
unop(Iop_32to8,
unop(Iop_V128to32, mkexpr(vB)))) );
putVReg( vD_addr,
binop(Iop_ShlV128, mkexpr(vA), mkexpr(sh)) );
break;
}
/* Shift Right */
case 0x204: // vsrb (Shift Right B, AV p256)
DIP("vsrb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shr8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x244: // vsrh (Shift Right HW, AV p257)
DIP("vsrh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shr16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x284: // vsrw (Shift Right W, AV p259)
DIP("vsrw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Shr32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x2C4: { // vsr (Shift Right, AV p251)
IRTemp sh = newTemp(Ity_I8);
DIP("vsr v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( sh, binop(Iop_And8, mkU8(0x7),
unop(Iop_32to8,
unop(Iop_V128to32, mkexpr(vB)))) );
putVReg( vD_addr,
binop(Iop_ShrV128, mkexpr(vA), mkexpr(sh)) );
break;
}
case 0x304: // vsrab (Shift Right Alg B, AV p253)
DIP("vsrab v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sar8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x344: // vsrah (Shift Right Alg HW, AV p254)
DIP("vsrah v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sar16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x384: // vsraw (Shift Right Alg W, AV p255)
DIP("vsraw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sar32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x44C: { // vsro (Shift Right by Octet, AV p258)
IRTemp sh = newTemp(Ity_I8);
DIP("vsro v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( sh, binop(Iop_And8, mkU8(0x78),
unop(Iop_32to8,
unop(Iop_V128to32, mkexpr(vB)))) );
putVReg( vD_addr,
binop(Iop_ShrV128, mkexpr(vA), mkexpr(sh)) );
break;
}
default:
vex_printf("dis_av_shift(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Permute Instructions
*/
static Bool dis_av_permute ( UInt theInstr )
{
/* VA-Form, VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar UIMM_5 = vA_addr;
UChar vB_addr = ifieldRegB(theInstr);
UChar vC_addr = ifieldRegC(theInstr);
UChar b10 = ifieldBIT10(theInstr);
UChar SHB_uimm4 = toUChar( IFIELD( theInstr, 6, 4 ) );
UInt opc2 = toUChar( IFIELD( theInstr, 0, 6 ) );
UChar SIMM_8 = extend_s_5to8(UIMM_5);
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp vC = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
assign( vC, getVReg(vC_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_permute(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x2A: // vsel (Conditional Select, AV p238)
DIP("vsel v%d,v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr, vC_addr);
/* vD = (vA & ~vC) | (vB & vC) */
putVReg( vD_addr, binop(Iop_OrV128,
binop(Iop_AndV128, mkexpr(vA), unop(Iop_NotV128, mkexpr(vC))),
binop(Iop_AndV128, mkexpr(vB), mkexpr(vC))) );
return True;
case 0x2B: { // vperm (Permute, AV p218)
/* limited to two args for IR, so have to play games... */
IRTemp a_perm = newTemp(Ity_V128);
IRTemp b_perm = newTemp(Ity_V128);
IRTemp mask = newTemp(Ity_V128);
IRTemp vC_andF = newTemp(Ity_V128);
DIP("vperm v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vB_addr, vC_addr);
/* Limit the Perm8x16 steering values to 0 .. 15 as that is what
IR specifies, and also to hide irrelevant bits from
memcheck */
assign( vC_andF,
binop(Iop_AndV128, mkexpr(vC),
unop(Iop_Dup8x16, mkU8(0xF))) );
assign( a_perm,
binop(Iop_Perm8x16, mkexpr(vA), mkexpr(vC_andF)) );
assign( b_perm,
binop(Iop_Perm8x16, mkexpr(vB), mkexpr(vC_andF)) );
// mask[i8] = (vC[i8]_4 == 1) ? 0xFF : 0x0
assign( mask, binop(Iop_SarN8x16,
binop(Iop_ShlN8x16, mkexpr(vC), mkU8(3)),
mkU8(7)) );
// dst = (a & ~mask) | (b & mask)
putVReg( vD_addr, binop(Iop_OrV128,
binop(Iop_AndV128, mkexpr(a_perm),
unop(Iop_NotV128, mkexpr(mask))),
binop(Iop_AndV128, mkexpr(b_perm),
mkexpr(mask))) );
return True;
}
case 0x2C: // vsldoi (Shift Left Double by Octet Imm, AV p241)
if (b10 != 0) {
vex_printf("dis_av_permute(ppc)(vsldoi)\n");
return False;
}
DIP("vsldoi v%d,v%d,v%d,%d\n",
vD_addr, vA_addr, vB_addr, SHB_uimm4);
if (SHB_uimm4 == 0)
putVReg( vD_addr, mkexpr(vA) );
else
putVReg( vD_addr,
binop(Iop_OrV128,
binop(Iop_ShlV128, mkexpr(vA), mkU8(SHB_uimm4*8)),
binop(Iop_ShrV128, mkexpr(vB), mkU8((16-SHB_uimm4)*8))) );
return True;
default:
break; // Fall through...
}
opc2 = IFIELD( theInstr, 0, 11 );
switch (opc2) {
/* Merge */
case 0x00C: // vmrghb (Merge High B, AV p195)
DIP("vmrghb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveHI8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x04C: // vmrghh (Merge High HW, AV p196)
DIP("vmrghh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveHI16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x08C: // vmrghw (Merge High W, AV p197)
DIP("vmrghw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveHI32x4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x10C: // vmrglb (Merge Low B, AV p198)
DIP("vmrglb v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveLO8x16, mkexpr(vA), mkexpr(vB)) );
break;
case 0x14C: // vmrglh (Merge Low HW, AV p199)
DIP("vmrglh v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveLO16x8, mkexpr(vA), mkexpr(vB)) );
break;
case 0x18C: // vmrglw (Merge Low W, AV p200)
DIP("vmrglw v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_InterleaveLO32x4, mkexpr(vA), mkexpr(vB)) );
break;
/* Splat */
case 0x20C: { // vspltb (Splat Byte, AV p245)
/* vD = Dup8x16( vB[UIMM_5] ) */
UChar sh_uimm = (15 - (UIMM_5 & 15)) * 8;
DIP("vspltb v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr, unop(Iop_Dup8x16,
unop(Iop_32to8, unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vB), mkU8(sh_uimm))))) );
break;
}
case 0x24C: { // vsplth (Splat Half Word, AV p246)
UChar sh_uimm = (7 - (UIMM_5 & 7)) * 16;
DIP("vsplth v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr, unop(Iop_Dup16x8,
unop(Iop_32to16, unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vB), mkU8(sh_uimm))))) );
break;
}
case 0x28C: { // vspltw (Splat Word, AV p250)
/* vD = Dup32x4( vB[UIMM_5] ) */
UChar sh_uimm = (3 - (UIMM_5 & 3)) * 32;
DIP("vspltw v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr, unop(Iop_Dup32x4,
unop(Iop_V128to32,
binop(Iop_ShrV128, mkexpr(vB), mkU8(sh_uimm)))) );
break;
}
case 0x30C: // vspltisb (Splat Immediate Signed B, AV p247)
DIP("vspltisb v%d,%d\n", vD_addr, (Char)SIMM_8);
putVReg( vD_addr, unop(Iop_Dup8x16, mkU8(SIMM_8)) );
break;
case 0x34C: // vspltish (Splat Immediate Signed HW, AV p248)
DIP("vspltish v%d,%d\n", vD_addr, (Char)SIMM_8);
putVReg( vD_addr,
unop(Iop_Dup16x8, mkU16(extend_s_8to32(SIMM_8))) );
break;
case 0x38C: // vspltisw (Splat Immediate Signed W, AV p249)
DIP("vspltisw v%d,%d\n", vD_addr, (Char)SIMM_8);
putVReg( vD_addr,
unop(Iop_Dup32x4, mkU32(extend_s_8to32(SIMM_8))) );
break;
default:
vex_printf("dis_av_permute(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Pack/Unpack Instructions
*/
static Bool dis_av_pack ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
IRTemp signs = IRTemp_INVALID;
IRTemp zeros = IRTemp_INVALID;
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_pack(ppc)(instr)\n");
return False;
}
switch (opc2) {
/* Packing */
case 0x00E: // vpkuhum (Pack Unsigned HW Unsigned Modulo, AV p224)
DIP("vpkuhum v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_NarrowBin16to8x16, mkexpr(vA), mkexpr(vB)) );
return True;
case 0x04E: // vpkuwum (Pack Unsigned W Unsigned Modulo, AV p226)
DIP("vpkuwum v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_NarrowBin32to16x8, mkexpr(vA), mkexpr(vB)) );
return True;
case 0x08E: // vpkuhus (Pack Unsigned HW Unsigned Saturate, AV p225)
DIP("vpkuhus v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_QNarrowBin16Uto8Ux16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
return True;
case 0x0CE: // vpkuwus (Pack Unsigned W Unsigned Saturate, AV p227)
DIP("vpkuwus v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_QNarrowBin32Uto16Ux8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
return True;
case 0x10E: { // vpkshus (Pack Signed HW Unsigned Saturate, AV p221)
// This insn does a signed->unsigned saturating conversion.
// Conversion done here, then uses unsigned->unsigned vpk insn:
// => UnsignedSaturatingNarrow( x & ~ (x >>s 15) )
IRTemp vA_tmp = newTemp(Ity_V128);
IRTemp vB_tmp = newTemp(Ity_V128);
DIP("vpkshus v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( vA_tmp, binop(Iop_AndV128, mkexpr(vA),
unop(Iop_NotV128,
binop(Iop_SarN16x8,
mkexpr(vA), mkU8(15)))) );
assign( vB_tmp, binop(Iop_AndV128, mkexpr(vB),
unop(Iop_NotV128,
binop(Iop_SarN16x8,
mkexpr(vB), mkU8(15)))) );
putVReg( vD_addr, binop(Iop_QNarrowBin16Uto8Ux16,
mkexpr(vA_tmp), mkexpr(vB_tmp)) );
// TODO: set VSCR[SAT]
return True;
}
case 0x14E: { // vpkswus (Pack Signed W Unsigned Saturate, AV p223)
// This insn does a signed->unsigned saturating conversion.
// Conversion done here, then uses unsigned->unsigned vpk insn:
// => UnsignedSaturatingNarrow( x & ~ (x >>s 31) )
IRTemp vA_tmp = newTemp(Ity_V128);
IRTemp vB_tmp = newTemp(Ity_V128);
DIP("vpkswus v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( vA_tmp, binop(Iop_AndV128, mkexpr(vA),
unop(Iop_NotV128,
binop(Iop_SarN32x4,
mkexpr(vA), mkU8(31)))) );
assign( vB_tmp, binop(Iop_AndV128, mkexpr(vB),
unop(Iop_NotV128,
binop(Iop_SarN32x4,
mkexpr(vB), mkU8(31)))) );
putVReg( vD_addr, binop(Iop_QNarrowBin32Uto16Ux8,
mkexpr(vA_tmp), mkexpr(vB_tmp)) );
// TODO: set VSCR[SAT]
return True;
}
case 0x18E: // vpkshss (Pack Signed HW Signed Saturate, AV p220)
DIP("vpkshss v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_QNarrowBin16Sto8Sx16, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
return True;
case 0x1CE: // vpkswss (Pack Signed W Signed Saturate, AV p222)
DIP("vpkswss v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_QNarrowBin32Sto16Sx8, mkexpr(vA), mkexpr(vB)) );
// TODO: set VSCR[SAT]
return True;
case 0x30E: { // vpkpx (Pack Pixel, AV p219)
/* CAB: Worth a new primop? */
/* Using shifts to compact pixel elements, then packing them */
IRTemp a1 = newTemp(Ity_V128);
IRTemp a2 = newTemp(Ity_V128);
IRTemp a3 = newTemp(Ity_V128);
IRTemp a_tmp = newTemp(Ity_V128);
IRTemp b1 = newTemp(Ity_V128);
IRTemp b2 = newTemp(Ity_V128);
IRTemp b3 = newTemp(Ity_V128);
IRTemp b_tmp = newTemp(Ity_V128);
DIP("vpkpx v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
assign( a1, binop(Iop_ShlN16x8,
binop(Iop_ShrN32x4, mkexpr(vA), mkU8(19)),
mkU8(10)) );
assign( a2, binop(Iop_ShlN16x8,
binop(Iop_ShrN16x8, mkexpr(vA), mkU8(11)),
mkU8(5)) );
assign( a3, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vA), mkU8(8)),
mkU8(11)) );
assign( a_tmp, binop(Iop_OrV128, mkexpr(a1),
binop(Iop_OrV128, mkexpr(a2), mkexpr(a3))) );
assign( b1, binop(Iop_ShlN16x8,
binop(Iop_ShrN32x4, mkexpr(vB), mkU8(19)),
mkU8(10)) );
assign( b2, binop(Iop_ShlN16x8,
binop(Iop_ShrN16x8, mkexpr(vB), mkU8(11)),
mkU8(5)) );
assign( b3, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vB), mkU8(8)),
mkU8(11)) );
assign( b_tmp, binop(Iop_OrV128, mkexpr(b1),
binop(Iop_OrV128, mkexpr(b2), mkexpr(b3))) );
putVReg( vD_addr, binop(Iop_NarrowBin32to16x8,
mkexpr(a_tmp), mkexpr(b_tmp)) );
return True;
}
default:
break; // Fall through...
}
if (vA_addr != 0) {
vex_printf("dis_av_pack(ppc)(vA_addr)\n");
return False;
}
signs = newTemp(Ity_V128);
zeros = newTemp(Ity_V128);
assign( zeros, unop(Iop_Dup32x4, mkU32(0)) );
switch (opc2) {
/* Unpacking */
case 0x20E: { // vupkhsb (Unpack High Signed B, AV p277)
DIP("vupkhsb v%d,v%d\n", vD_addr, vB_addr);
assign( signs, binop(Iop_CmpGT8Sx16, mkexpr(zeros), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_InterleaveHI8x16, mkexpr(signs), mkexpr(vB)) );
break;
}
case 0x24E: { // vupkhsh (Unpack High Signed HW, AV p278)
DIP("vupkhsh v%d,v%d\n", vD_addr, vB_addr);
assign( signs, binop(Iop_CmpGT16Sx8, mkexpr(zeros), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_InterleaveHI16x8, mkexpr(signs), mkexpr(vB)) );
break;
}
case 0x28E: { // vupklsb (Unpack Low Signed B, AV p280)
DIP("vupklsb v%d,v%d\n", vD_addr, vB_addr);
assign( signs, binop(Iop_CmpGT8Sx16, mkexpr(zeros), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_InterleaveLO8x16, mkexpr(signs), mkexpr(vB)) );
break;
}
case 0x2CE: { // vupklsh (Unpack Low Signed HW, AV p281)
DIP("vupklsh v%d,v%d\n", vD_addr, vB_addr);
assign( signs, binop(Iop_CmpGT16Sx8, mkexpr(zeros), mkexpr(vB)) );
putVReg( vD_addr,
binop(Iop_InterleaveLO16x8, mkexpr(signs), mkexpr(vB)) );
break;
}
case 0x34E: { // vupkhpx (Unpack High Pixel16, AV p276)
/* CAB: Worth a new primop? */
/* Using shifts to isolate pixel elements, then expanding them */
IRTemp z0 = newTemp(Ity_V128);
IRTemp z1 = newTemp(Ity_V128);
IRTemp z01 = newTemp(Ity_V128);
IRTemp z2 = newTemp(Ity_V128);
IRTemp z3 = newTemp(Ity_V128);
IRTemp z23 = newTemp(Ity_V128);
DIP("vupkhpx v%d,v%d\n", vD_addr, vB_addr);
assign( z0, binop(Iop_ShlN16x8,
binop(Iop_SarN16x8, mkexpr(vB), mkU8(15)),
mkU8(8)) );
assign( z1, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vB), mkU8(1)),
mkU8(11)) );
assign( z01, binop(Iop_InterleaveHI16x8, mkexpr(zeros),
binop(Iop_OrV128, mkexpr(z0), mkexpr(z1))) );
assign( z2, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8,
binop(Iop_ShrN16x8, mkexpr(vB), mkU8(5)),
mkU8(11)),
mkU8(3)) );
assign( z3, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vB), mkU8(11)),
mkU8(11)) );
assign( z23, binop(Iop_InterleaveHI16x8, mkexpr(zeros),
binop(Iop_OrV128, mkexpr(z2), mkexpr(z3))) );
putVReg( vD_addr,
binop(Iop_OrV128,
binop(Iop_ShlN32x4, mkexpr(z01), mkU8(16)),
mkexpr(z23)) );
break;
}
case 0x3CE: { // vupklpx (Unpack Low Pixel16, AV p279)
/* identical to vupkhpx, except interleaving LO */
IRTemp z0 = newTemp(Ity_V128);
IRTemp z1 = newTemp(Ity_V128);
IRTemp z01 = newTemp(Ity_V128);
IRTemp z2 = newTemp(Ity_V128);
IRTemp z3 = newTemp(Ity_V128);
IRTemp z23 = newTemp(Ity_V128);
DIP("vupklpx v%d,v%d\n", vD_addr, vB_addr);
assign( z0, binop(Iop_ShlN16x8,
binop(Iop_SarN16x8, mkexpr(vB), mkU8(15)),
mkU8(8)) );
assign( z1, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vB), mkU8(1)),
mkU8(11)) );
assign( z01, binop(Iop_InterleaveLO16x8, mkexpr(zeros),
binop(Iop_OrV128, mkexpr(z0), mkexpr(z1))) );
assign( z2, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8,
binop(Iop_ShrN16x8, mkexpr(vB), mkU8(5)),
mkU8(11)),
mkU8(3)) );
assign( z3, binop(Iop_ShrN16x8,
binop(Iop_ShlN16x8, mkexpr(vB), mkU8(11)),
mkU8(11)) );
assign( z23, binop(Iop_InterleaveLO16x8, mkexpr(zeros),
binop(Iop_OrV128, mkexpr(z2), mkexpr(z3))) );
putVReg( vD_addr,
binop(Iop_OrV128,
binop(Iop_ShlN32x4, mkexpr(z01), mkU8(16)),
mkexpr(z23)) );
break;
}
default:
vex_printf("dis_av_pack(ppc)(opc2)\n");
return False;
}
return True;
}
/*
AltiVec Floating Point Arithmetic Instructions
*/
static Bool dis_av_fp_arith ( UInt theInstr )
{
/* VA-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UChar vC_addr = ifieldRegC(theInstr);
UInt opc2=0;
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp vC = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
assign( vC, getVReg(vC_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_fp_arith(ppc)(instr)\n");
return False;
}
opc2 = IFIELD( theInstr, 0, 6 );
switch (opc2) {
case 0x2E: // vmaddfp (Multiply Add FP, AV p177)
DIP("vmaddfp v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vC_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_Add32Fx4, mkexpr(vB),
binop(Iop_Mul32Fx4, mkexpr(vA), mkexpr(vC))) );
return True;
case 0x2F: { // vnmsubfp (Negative Multiply-Subtract FP, AV p215)
DIP("vnmsubfp v%d,v%d,v%d,v%d\n",
vD_addr, vA_addr, vC_addr, vB_addr);
putVReg( vD_addr,
binop(Iop_Sub32Fx4,
mkexpr(vB),
binop(Iop_Mul32Fx4, mkexpr(vA), mkexpr(vC))) );
return True;
}
default:
break; // Fall through...
}
opc2 = IFIELD( theInstr, 0, 11 );
switch (opc2) {
case 0x00A: // vaddfp (Add FP, AV p137)
DIP("vaddfp v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Add32Fx4, mkexpr(vA), mkexpr(vB)) );
return True;
case 0x04A: // vsubfp (Subtract FP, AV p261)
DIP("vsubfp v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Sub32Fx4, mkexpr(vA), mkexpr(vB)) );
return True;
case 0x40A: // vmaxfp (Maximum FP, AV p178)
DIP("vmaxfp v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Max32Fx4, mkexpr(vA), mkexpr(vB)) );
return True;
case 0x44A: // vminfp (Minimum FP, AV p187)
DIP("vminfp v%d,v%d,v%d\n", vD_addr, vA_addr, vB_addr);
putVReg( vD_addr, binop(Iop_Min32Fx4, mkexpr(vA), mkexpr(vB)) );
return True;
default:
break; // Fall through...
}
if (vA_addr != 0) {
vex_printf("dis_av_fp_arith(ppc)(vA_addr)\n");
return False;
}
switch (opc2) {
case 0x10A: // vrefp (Reciprocal Esimate FP, AV p228)
DIP("vrefp v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_Recip32Fx4, mkexpr(vB)) );
return True;
case 0x14A: // vrsqrtefp (Reciprocal Sqrt Estimate FP, AV p237)
DIP("vrsqrtefp v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_RSqrt32Fx4, mkexpr(vB)) );
return True;
case 0x18A: // vexptefp (2 Raised to the Exp Est FP, AV p173)
DIP("vexptefp v%d,v%d\n", vD_addr, vB_addr);
DIP(" => not implemented\n");
return False;
case 0x1CA: // vlogefp (Log2 Estimate FP, AV p175)
DIP("vlogefp v%d,v%d\n", vD_addr, vB_addr);
DIP(" => not implemented\n");
return False;
default:
vex_printf("dis_av_fp_arith(ppc)(opc2=0x%x)\n",opc2);
return False;
}
return True;
}
/*
AltiVec Floating Point Compare Instructions
*/
static Bool dis_av_fp_cmp ( UInt theInstr )
{
/* VXR-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar vA_addr = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UChar flag_rC = ifieldBIT10(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 10 );
Bool cmp_bounds = False;
IRTemp vA = newTemp(Ity_V128);
IRTemp vB = newTemp(Ity_V128);
IRTemp vD = newTemp(Ity_V128);
assign( vA, getVReg(vA_addr));
assign( vB, getVReg(vB_addr));
if (opc1 != 0x4) {
vex_printf("dis_av_fp_cmp(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x0C6: // vcmpeqfp (Compare Equal-to FP, AV p159)
DIP("vcmpeqfp%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpEQ32Fx4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x1C6: // vcmpgefp (Compare Greater-than-or-Equal-to, AV p163)
DIP("vcmpgefp%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGE32Fx4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x2C6: // vcmpgtfp (Compare Greater-than FP, AV p164)
DIP("vcmpgtfp%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
assign( vD, binop(Iop_CmpGT32Fx4, mkexpr(vA), mkexpr(vB)) );
break;
case 0x3C6: { // vcmpbfp (Compare Bounds FP, AV p157)
IRTemp gt = newTemp(Ity_V128);
IRTemp lt = newTemp(Ity_V128);
IRTemp zeros = newTemp(Ity_V128);
DIP("vcmpbfp%s v%d,v%d,v%d\n", (flag_rC ? ".":""),
vD_addr, vA_addr, vB_addr);
cmp_bounds = True;
assign( zeros, unop(Iop_Dup32x4, mkU32(0)) );
/* Note: making use of fact that the ppc backend for compare insns
return zero'd lanes if either of the corresponding arg lanes is
a nan.
Perhaps better to have an irop Iop_isNan32Fx4, but then we'd
need this for the other compares too (vcmpeqfp etc)...
Better still, tighten down the spec for compare irops.
*/
assign( gt, unop(Iop_NotV128,
binop(Iop_CmpLE32Fx4, mkexpr(vA), mkexpr(vB))) );
assign( lt, unop(Iop_NotV128,
binop(Iop_CmpGE32Fx4, mkexpr(vA),
binop(Iop_Sub32Fx4, mkexpr(zeros),
mkexpr(vB)))) );
// finally, just shift gt,lt to correct position
assign( vD, binop(Iop_ShlN32x4,
binop(Iop_OrV128,
binop(Iop_AndV128, mkexpr(gt),
unop(Iop_Dup32x4, mkU32(0x2))),
binop(Iop_AndV128, mkexpr(lt),
unop(Iop_Dup32x4, mkU32(0x1)))),
mkU8(30)) );
break;
}
default:
vex_printf("dis_av_fp_cmp(ppc)(opc2)\n");
return False;
}
putVReg( vD_addr, mkexpr(vD) );
if (flag_rC) {
set_AV_CR6( mkexpr(vD), !cmp_bounds );
}
return True;
}
/*
AltiVec Floating Point Convert/Round Instructions
*/
static Bool dis_av_fp_convert ( UInt theInstr )
{
/* VX-Form */
UChar opc1 = ifieldOPC(theInstr);
UChar vD_addr = ifieldRegDS(theInstr);
UChar UIMM_5 = ifieldRegA(theInstr);
UChar vB_addr = ifieldRegB(theInstr);
UInt opc2 = IFIELD( theInstr, 0, 11 );
IRTemp vB = newTemp(Ity_V128);
IRTemp vScale = newTemp(Ity_V128);
IRTemp vInvScale = newTemp(Ity_V128);
float scale, inv_scale;
assign( vB, getVReg(vB_addr));
/* scale = 2^UIMM, cast to float, reinterpreted as uint */
scale = (float)( (unsigned int) 1<<UIMM_5 );
assign( vScale, unop(Iop_Dup32x4, mkU32( float_to_bits(scale) )) );
inv_scale = 1/scale;
assign( vInvScale,
unop(Iop_Dup32x4, mkU32( float_to_bits(inv_scale) )) );
if (opc1 != 0x4) {
vex_printf("dis_av_fp_convert(ppc)(instr)\n");
return False;
}
switch (opc2) {
case 0x30A: // vcfux (Convert from Unsigned Fixed-Point W, AV p156)
DIP("vcfux v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr, binop(Iop_Mul32Fx4,
unop(Iop_I32UtoFx4, mkexpr(vB)),
mkexpr(vInvScale)) );
return True;
case 0x34A: // vcfsx (Convert from Signed Fixed-Point W, AV p155)
DIP("vcfsx v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr, binop(Iop_Mul32Fx4,
unop(Iop_I32StoFx4, mkexpr(vB)),
mkexpr(vInvScale)) );
return True;
case 0x38A: // vctuxs (Convert to Unsigned Fixed-Point W Saturate, AV p172)
DIP("vctuxs v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr,
unop(Iop_QFtoI32Ux4_RZ,
binop(Iop_Mul32Fx4, mkexpr(vB), mkexpr(vScale))) );
return True;
case 0x3CA: // vctsxs (Convert to Signed Fixed-Point W Saturate, AV p171)
DIP("vctsxs v%d,v%d,%d\n", vD_addr, vB_addr, UIMM_5);
putVReg( vD_addr,
unop(Iop_QFtoI32Sx4_RZ,
binop(Iop_Mul32Fx4, mkexpr(vB), mkexpr(vScale))) );
return True;
default:
break; // Fall through...
}
if (UIMM_5 != 0) {
vex_printf("dis_av_fp_convert(ppc)(UIMM_5)\n");
return False;
}
switch (opc2) {
case 0x20A: // vrfin (Round to FP Integer Nearest, AV p231)
DIP("vrfin v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_RoundF32x4_RN, mkexpr(vB)) );
break;
case 0x24A: // vrfiz (Round to FP Integer toward zero, AV p233)
DIP("vrfiz v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_RoundF32x4_RZ, mkexpr(vB)) );
break;
case 0x28A: // vrfip (Round to FP Integer toward +inf, AV p232)
DIP("vrfip v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_RoundF32x4_RP, mkexpr(vB)) );
break;
case 0x2CA: // vrfim (Round to FP Integer toward -inf, AV p230)
DIP("vrfim v%d,v%d\n", vD_addr, vB_addr);
putVReg( vD_addr, unop(Iop_RoundF32x4_RM, mkexpr(vB)) );
break;
default:
vex_printf("dis_av_fp_convert(ppc)(opc2)\n");
return False;
}
return True;
}
/* The 0x3C primary opcode (VSX category) uses several different forms of
* extended opcodes:
* o XX2-form:
* - [10:2] (IBM notation [21:29])
* o XX3-form variants:
* - variant 1: [10:3] (IBM notation [21:28])
* - variant 2: [9:3] (IBM notation [22:28])
* - variant 3: [7:3] (IBM notation [24:28])
* o XX-4 form:
* - [10:6] (IBM notation [21:25])
*
* The XX2-form needs bit 0 masked from the standard extended opcode
* as returned by ifieldOPClo10; the XX3-form needs bits 0 and 1 masked;
* and the XX4-form needs bits 0, 1, and 2 masked. Additionally, the
* XX4 and XX3 (variants 2 and 3) forms need certain bits masked on the
* front end since their encoding does not begin at bit 21 like the standard
* format.
*
* The get_VSX60_opc2() function uses the vsx_insn array below to obtain the
* secondary opcode for such VSX instructions.
*
*/
struct vsx_insn {
UInt opcode;
Char * name;
};
// ATTENTION: Keep this array sorted on the opcocde!!!
static struct vsx_insn vsx_all[] = {
{ 0x8, "xxsldwi" },
{ 0x18, "xxsel" },
{ 0x28, "xxpermdi" },
{ 0x48, "xxmrghw" },
{ 0x80, "xsadddp" },
{ 0x84, "xsmaddadp" },
{ 0x8c, "xscmpudp" },
{ 0x90, "xscvdpuxws" },
{ 0x92, "xsrdpi" },
{ 0x94, "xsrsqrtedp" },
{ 0x96, "xssqrtdp" },
{ 0xa0, "xssubdp" },
{ 0xa4, "xsmaddmdp" },
{ 0xac, "xscmpodp" },
{ 0xb0, "xscvdpsxws" },
{ 0xb2, "xsrdpiz" },
{ 0xb4, "xsredp" },
{ 0xc0, "xsmuldp" },
{ 0xc4, "xsmsubadp" },
{ 0xc8, "xxmrglw" },
{ 0xd2, "xsrdpip" },
{ 0xd4, "xstsqrtdp" },
{ 0xd6, "xsrdpic" },
{ 0xe0, "xsdivdp" },
{ 0xe4, "xsmsubmdp" },
{ 0xf2, "xsrdpim" },
{ 0xf4, "xstdivdp" },
{ 0x100, "xvaddsp" },
{ 0x104, "xvmaddasp" },
{ 0x10c, "xvcmpeqsp" },
{ 0x110, "xvcvspuxws" },
{ 0x112, "xvrspi" },
{ 0x114, "xvrsqrtesp" },
{ 0x116, "xvsqrtsp" },
{ 0x120, "xvsubsp" },
{ 0x124, "xvmaddmsp" },
{ 0x12c, "xvcmpgtsp" },
{ 0x130, "xvcvspsxws" },
{ 0x132, "xvrspiz" },
{ 0x134, "xvresp" },
{ 0x140, "xvmulsp" },
{ 0x144, "xvmsubasp" },
{ 0x148, "xxspltw" },
{ 0x14c, "xvcmpgesp" },
{ 0x150, "xvcvuxwsp" },
{ 0x152, "xvrspip" },
{ 0x154, "xvtsqrtsp" },
{ 0x156, "xvrspic" },
{ 0x160, "xvdivsp" },
{ 0x164, "xvmsubmsp" },
{ 0x170, "xvcvsxwsp" },
{ 0x172, "xvrspim" },
{ 0x174, "xvtdivsp" },
{ 0x180, "xvadddp" },
{ 0x184, "xvmaddadp" },
{ 0x18c, "xvcmpeqdp" },
{ 0x190, "xvcvdpuxws" },
{ 0x192, "xvrdpi" },
{ 0x194, "xvrsqrtedp" },
{ 0x196, "xvsqrtdp" },
{ 0x1a0, "xvsubdp" },
{ 0x1a4, "xvmaddmdp" },
{ 0x1ac, "xvcmpgtdp" },
{ 0x1b0, "xvcvdpsxws" },
{ 0x1b2, "xvrdpiz" },
{ 0x1b4, "xvredp" },
{ 0x1c0, "xvmuldp" },
{ 0x1c4, "xvmsubadp" },
{ 0x1cc, "xvcmpgedp" },
{ 0x1d0, "xvcvuxwdp" },
{ 0x1d2, "xvrdpip" },
{ 0x1d4, "xvtsqrtdp" },
{ 0x1d6, "xvrdpic" },
{ 0x1e0, "xvdivdp" },
{ 0x1e4, "xvmsubmdp" },
{ 0x1f0, "xvcvsxwdp" },
{ 0x1f2, "xvrdpim" },
{ 0x1f4, "xvtdivdp" },
{ 0x208, "xxland" },
{ 0x212, "xscvdpsp" },
{ 0x228, "xxlandc" },
{ 0x248 , "xxlor" },
{ 0x268, "xxlxor" },
{ 0x280, "xsmaxdp" },
{ 0x284, "xsnmaddadp" },
{ 0x288, "xxlnor" },
{ 0x290, "xscvdpuxds" },
{ 0x292, "xscvspdp" },
{ 0x2a0, "xsmindp" },
{ 0x2a4, "xsnmaddmdp" },
{ 0x2b0, "xscvdpsxds" },
{ 0x2b2, "xsabsdp" },
{ 0x2c0, "xscpsgndp" },
{ 0x2c4, "xsnmsubadp" },
{ 0x2d0, "xscvuxddp" },
{ 0x2d2, "xsnabsdp" },
{ 0x2e4, "xsnmsubmdp" },
{ 0x2f0, "xscvsxddp" },
{ 0x2f2, "xsnegdp" },
{ 0x300, "xvmaxsp" },
{ 0x304, "xvnmaddasp" },
{ 0x30c, "xvcmpeqsp." },
{ 0x310, "xvcvspuxds" },
{ 0x312, "xvcvdpsp" },
{ 0x320, "xvminsp" },
{ 0x324, "xvnmaddmsp" },
{ 0x32c, "xvcmpgtsp." },
{ 0x330, "xvcvspsxds" },
{ 0x332, "xvabssp" },
{ 0x340, "xvcpsgnsp" },
{ 0x344, "xvnmsubasp" },
{ 0x34c, "xvcmpgesp." },
{ 0x350, "xvcvuxdsp" },
{ 0x352, "xvnabssp" },
{ 0x364, "xvnmsubmsp" },
{ 0x370, "xvcvsxdsp" },
{ 0x372, "xvnegsp" },
{ 0x380, "xvmaxdp" },
{ 0x384, "xvnmaddadp" },
{ 0x38c, "xvcmpeqdp." },
{ 0x390, "xvcvdpuxds" },
{ 0x392, "xvcvspdp" },
{ 0x3a0, "xvmindp" },
{ 0x3a4, "xvnmaddmdp" },
{ 0x3ac, "xvcmpgtdp." },
{ 0x3b0, "xvcvdpsxds" },
{ 0x3b2, "xvabsdp" },
{ 0x3c0, "xvcpsgndp" },
{ 0x3c4, "xvnmsubadp" },
{ 0x3cc, "xvcmpgedp." },
{ 0x3d0, "xvcvuxddp" },
{ 0x3d2, "xvnabsdp" },
{ 0x3e4, "xvnmsubmdp" },
{ 0x3f0, "xvcvsxddp" },
{ 0x3f2, "xvnegdp" }
};
#define VSX_ALL_LEN 135
// ATTENTION: This search function assumes vsx_all array is sorted.
static Int findVSXextOpCode(UInt opcode)
{
Int low, mid, high;
low = 0;
high = VSX_ALL_LEN - 1;
while (low <= high) {
mid = (low + high)/2;
if (opcode < vsx_all[mid].opcode)
high = mid - 1;
else if (opcode > vsx_all[mid].opcode)
low = mid + 1;
else
return mid;
}
return -1;
}
/* The full 10-bit extended opcode retrieved via ifieldOPClo10 is
* passed, and we then try to match it up with one of the VSX forms
* below.
*/
static UInt get_VSX60_opc2(UInt opc2_full)
{
#define XX2_MASK 0x000003FE
#define XX3_1_MASK 0x000003FC
#define XX3_2_MASK 0x000001FC
#define XX3_3_MASK 0x0000007C
#define XX4_MASK 0x00000018
Int ret;
UInt vsxExtOpcode = 0;
if (( ret = findVSXextOpCode(opc2_full & XX2_MASK)) >= 0)
vsxExtOpcode = vsx_all[ret].opcode;
else if (( ret = findVSXextOpCode(opc2_full & XX3_1_MASK)) >= 0)
vsxExtOpcode = vsx_all[ret].opcode;
else if (( ret = findVSXextOpCode(opc2_full & XX3_2_MASK)) >= 0)
vsxExtOpcode = vsx_all[ret].opcode;
else if (( ret = findVSXextOpCode(opc2_full & XX3_3_MASK)) >= 0)
vsxExtOpcode = vsx_all[ret].opcode;
else if (( ret = findVSXextOpCode(opc2_full & XX4_MASK)) >= 0)
vsxExtOpcode = vsx_all[ret].opcode;
return vsxExtOpcode;
}
/*------------------------------------------------------------*/
/*--- Disassemble a single instruction ---*/
/*------------------------------------------------------------*/
/* Disassemble a single instruction into IR. The instruction
is located in host memory at &guest_code[delta]. */
static
DisResult disInstr_PPC_WRK (
Bool (*resteerOkFn) ( /*opaque*/void*, Addr64 ),
Bool resteerCisOk,
void* callback_opaque,
Long delta64,
VexArchInfo* archinfo,
VexAbiInfo* abiinfo
)
{
UChar opc1;
UInt opc2;
DisResult dres;
UInt theInstr;
IRType ty = mode64 ? Ity_I64 : Ity_I32;
Bool allow_F = False;
Bool allow_V = False;
Bool allow_FX = False;
Bool allow_GX = False;
Bool allow_VX = False; // Equates to "supports Power ISA 2.06
Bool allow_DFP = False;
UInt hwcaps = archinfo->hwcaps;
Long delta;
/* What insn variants are we supporting today? */
if (mode64) {
allow_F = True;
allow_V = (0 != (hwcaps & VEX_HWCAPS_PPC64_V));
allow_FX = (0 != (hwcaps & VEX_HWCAPS_PPC64_FX));
allow_GX = (0 != (hwcaps & VEX_HWCAPS_PPC64_GX));
allow_VX = (0 != (hwcaps & VEX_HWCAPS_PPC64_VX));
allow_DFP = (0 != (hwcaps & VEX_HWCAPS_PPC64_DFP));
} else {
allow_F = (0 != (hwcaps & VEX_HWCAPS_PPC32_F));
allow_V = (0 != (hwcaps & VEX_HWCAPS_PPC32_V));
allow_FX = (0 != (hwcaps & VEX_HWCAPS_PPC32_FX));
allow_GX = (0 != (hwcaps & VEX_HWCAPS_PPC32_GX));
allow_VX = (0 != (hwcaps & VEX_HWCAPS_PPC32_VX));
allow_DFP = (0 != (hwcaps & VEX_HWCAPS_PPC32_DFP));
}
/* The running delta */
delta = (Long)mkSzAddr(ty, (ULong)delta64);
/* Set result defaults. */
dres.whatNext = Dis_Continue;
dres.len = 0;
dres.continueAt = 0;
dres.jk_StopHere = Ijk_INVALID;
/* At least this is simple on PPC32: insns are all 4 bytes long, and
4-aligned. So just fish the whole thing out of memory right now
and have done. */
theInstr = getUIntBigendianly( (UChar*)(&guest_code[delta]) );
if (0) vex_printf("insn: 0x%x\n", theInstr);
DIP("\t0x%llx: ", (ULong)guest_CIA_curr_instr);
/* Spot "Special" instructions (see comment at top of file). */
{
UChar* code = (UChar*)(guest_code + delta);
/* Spot the 16-byte preamble:
32-bit mode:
54001800 rlwinm 0,0,3,0,0
54006800 rlwinm 0,0,13,0,0
5400E800 rlwinm 0,0,29,0,0
54009800 rlwinm 0,0,19,0,0
64-bit mode:
78001800 rotldi 0,0,3
78006800 rotldi 0,0,13
7800E802 rotldi 0,0,61
78009802 rotldi 0,0,51
*/
UInt word1 = mode64 ? 0x78001800 : 0x54001800;
UInt word2 = mode64 ? 0x78006800 : 0x54006800;
UInt word3 = mode64 ? 0x7800E802 : 0x5400E800;
UInt word4 = mode64 ? 0x78009802 : 0x54009800;
if (getUIntBigendianly(code+ 0) == word1 &&
getUIntBigendianly(code+ 4) == word2 &&
getUIntBigendianly(code+ 8) == word3 &&
getUIntBigendianly(code+12) == word4) {
/* Got a "Special" instruction preamble. Which one is it? */
if (getUIntBigendianly(code+16) == 0x7C210B78 /* or 1,1,1 */) {
/* %R3 = client_request ( %R4 ) */
DIP("r3 = client_request ( %%r4 )\n");
delta += 20;
putGST( PPC_GST_CIA, mkSzImm( ty, guest_CIA_bbstart + delta ));
dres.jk_StopHere = Ijk_ClientReq;
dres.whatNext = Dis_StopHere;
goto decode_success;
}
else
if (getUIntBigendianly(code+16) == 0x7C421378 /* or 2,2,2 */) {
/* %R3 = guest_NRADDR */
DIP("r3 = guest_NRADDR\n");
delta += 20;
dres.len = 20;
putIReg(3, IRExpr_Get( OFFB_NRADDR, ty ));
goto decode_success;
}
else
if (getUIntBigendianly(code+16) == 0x7C631B78 /* or 3,3,3 */) {
/* branch-and-link-to-noredir %R11 */
DIP("branch-and-link-to-noredir r11\n");
delta += 20;
putGST( PPC_GST_LR, mkSzImm(ty, guest_CIA_bbstart + (Long)delta) );
putGST( PPC_GST_CIA, getIReg(11));
dres.jk_StopHere = Ijk_NoRedir;
dres.whatNext = Dis_StopHere;
goto decode_success;
}
else
if (getUIntBigendianly(code+16) == 0x7C842378 /* or 4,4,4 */) {
/* %R3 = guest_NRADDR_GPR2 */
DIP("r3 = guest_NRADDR_GPR2\n");
delta += 20;
dres.len = 20;
putIReg(3, IRExpr_Get( OFFB_NRADDR_GPR2, ty ));
goto decode_success;
}
/* We don't know what it is. Set opc1/opc2 so decode_failure
can print the insn following the Special-insn preamble. */
theInstr = getUIntBigendianly(code+16);
opc1 = ifieldOPC(theInstr);
opc2 = ifieldOPClo10(theInstr);
goto decode_failure;
/*NOTREACHED*/
}
}
opc1 = ifieldOPC(theInstr);
opc2 = ifieldOPClo10(theInstr);
// Note: all 'reserved' bits must be cleared, else invalid
switch (opc1) {
/* Integer Arithmetic Instructions */
case 0x0C: case 0x0D: case 0x0E: // addic, addic., addi
case 0x0F: case 0x07: case 0x08: // addis, mulli, subfic
if (dis_int_arith( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Compare Instructions */
case 0x0B: case 0x0A: // cmpi, cmpli
if (dis_int_cmp( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Logical Instructions */
case 0x1C: case 0x1D: case 0x18: // andi., andis., ori
case 0x19: case 0x1A: case 0x1B: // oris, xori, xoris
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Rotate Instructions */
case 0x14: case 0x15: case 0x17: // rlwimi, rlwinm, rlwnm
if (dis_int_rot( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Rotate Instructions */
case 0x1E: // rldcl, rldcr, rldic, rldicl, rldicr, rldimi
if (dis_int_rot( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Load Instructions */
case 0x22: case 0x23: case 0x2A: // lbz, lbzu, lha
case 0x2B: case 0x28: case 0x29: // lhau, lhz, lhzu
case 0x20: case 0x21: // lwz, lwzu
if (dis_int_load( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Store Instructions */
case 0x26: case 0x27: case 0x2C: // stb, stbu, sth
case 0x2D: case 0x24: case 0x25: // sthu, stw, stwu
if (dis_int_store( theInstr, abiinfo )) goto decode_success;
goto decode_failure;
/* Integer Load and Store Multiple Instructions */
case 0x2E: case 0x2F: // lmw, stmw
if (dis_int_ldst_mult( theInstr )) goto decode_success;
goto decode_failure;
/* Branch Instructions */
case 0x12: case 0x10: // b, bc
if (dis_branch(theInstr, abiinfo, &dres,
resteerOkFn, callback_opaque))
goto decode_success;
goto decode_failure;
/* System Linkage Instructions */
case 0x11: // sc
if (dis_syslink(theInstr, abiinfo, &dres)) goto decode_success;
goto decode_failure;
/* Trap Instructions */
case 0x02: case 0x03: // tdi, twi
if (dis_trapi(theInstr, &dres)) goto decode_success;
goto decode_failure;
/* Floating Point Load Instructions */
case 0x30: case 0x31: case 0x32: // lfs, lfsu, lfd
case 0x33: // lfdu
if (!allow_F) goto decode_noF;
if (dis_fp_load( theInstr )) goto decode_success;
goto decode_failure;
/* Floating Point Store Instructions */
case 0x34: case 0x35: case 0x36: // stfsx, stfsux, stfdx
case 0x37: // stfdux
if (!allow_F) goto decode_noF;
if (dis_fp_store( theInstr )) goto decode_success;
goto decode_failure;
/* Floating Point Load Double Pair Instructions */
case 0x39: case 0x3D:
if (!allow_F) goto decode_noF;
if (dis_fp_pair( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Loads */
case 0x3A: // ld, ldu, lwa
if (!mode64) goto decode_failure;
if (dis_int_load( theInstr )) goto decode_success;
goto decode_failure;
case 0x3B:
if (!allow_F) goto decode_noF;
opc2 = ifieldOPClo10(theInstr);
switch (opc2) {
case 0x2: // dadd - DFP Add
case 0x202: // dsub - DFP Subtract
case 0x22: // dmul - DFP Mult
case 0x222: // ddiv - DFP Divide
if (!allow_DFP) goto decode_noDFP;
if (dis_dfp_arith( theInstr ))
goto decode_success;
case 0x82: // dcmpo, DFP comparison ordered instruction
case 0x282: // dcmpu, DFP comparison unordered instruction
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_compare( theInstr ) )
goto decode_success;
goto decode_failure;
case 0x102: // dctdp - DFP convert to DFP long
case 0x302: // drsp - DFP round to dfp short
case 0x122: // dctfix - DFP convert to fixed
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_fmt_conv( theInstr ))
goto decode_success;
goto decode_failure;
case 0x322: // POWER 7 inst, dcffix - DFP convert from fixed
if (!allow_VX)
goto decode_failure;
if (dis_dfp_fmt_conv( theInstr ))
goto decode_success;
goto decode_failure;
case 0x2A2: // dtstsf - DFP number of significant digits
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_significant_digits(theInstr))
goto decode_success;
goto decode_failure;
case 0x142: // ddedpd DFP Decode DPD to BCD
case 0x342: // denbcd DFP Encode BCD to DPD
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_bcd(theInstr))
goto decode_success;
goto decode_failure;
case 0x162: // dxex - Extract exponent
case 0x362: // diex - Insert exponent
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_extract_insert( theInstr ) )
goto decode_success;
goto decode_failure;
case 0x3CE: // fcfidus (implemented as native insn)
if (!allow_VX)
goto decode_noVX;
if (dis_fp_round( theInstr ))
goto decode_success;
goto decode_failure;
case 0x34E: // fcfids
if (dis_fp_round( theInstr ))
goto decode_success;
goto decode_failure;
}
opc2 = ifieldOPClo9( theInstr );
switch (opc2) {
case 0x42: // dscli, DFP shift left
case 0x62: // dscri, DFP shift right
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_shift( theInstr ))
goto decode_success;
goto decode_failure;
case 0xc2: // dtstdc, DFP test data class
case 0xe2: // dtstdg, DFP test data group
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_class_test( theInstr ))
goto decode_success;
goto decode_failure;
}
opc2 = ifieldOPClo8( theInstr );
switch (opc2) {
case 0x3: // dqua - DFP Quantize
case 0x23: // drrnd - DFP Reround
case 0x43: // dquai - DFP Quantize immediate
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_quantize_sig_rrnd( theInstr ) )
goto decode_success;
goto decode_failure;
case 0xA2: // dtstex - DFP Test exponent
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_exponent_test( theInstr ) )
goto decode_success;
goto decode_failure;
case 0x63: // drintx - Round to an integer value
case 0xE3: // drintn - Round to an integer value
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_round( theInstr ) ) {
goto decode_success;
}
goto decode_failure;
default:
break; /* fall through to next opc2 check */
}
opc2 = IFIELD(theInstr, 1, 5);
switch (opc2) {
/* Floating Point Arith Instructions */
case 0x12: case 0x14: case 0x15: // fdivs, fsubs, fadds
case 0x19: // fmuls
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
case 0x16: // fsqrts
if (!allow_FX) goto decode_noFX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
case 0x18: // fres
if (!allow_GX) goto decode_noGX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
/* Floating Point Mult-Add Instructions */
case 0x1C: case 0x1D: case 0x1E: // fmsubs, fmadds, fnmsubs
case 0x1F: // fnmadds
if (dis_fp_multadd(theInstr)) goto decode_success;
goto decode_failure;
case 0x1A: // frsqrtes
if (!allow_GX) goto decode_noGX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
default:
goto decode_failure;
}
break;
case 0x3C: // VSX instructions (except load/store)
{
// All of these VSX instructions use some VMX facilities, so
// if allow_V is not set, we'll skip trying to decode.
if (!allow_V) goto decode_noVX;
UInt vsxOpc2 = get_VSX60_opc2(opc2);
/* The vsxOpc2 returned is the "normalized" value, representing the
* instructions secondary opcode as taken from the standard secondary
* opcode field [21:30] (IBM notatition), even if the actual field
* is non-standard. These normalized values are given in the opcode
* appendices of the ISA 2.06 document.
*/
if (vsxOpc2 == 0)
goto decode_failure;
switch (vsxOpc2) {
case 0x8: case 0x28: case 0x48: case 0xc8: // xxsldwi, xxpermdi, xxmrghw, xxmrglw
case 0x018: case 0x148: // xxsel, xxspltw
if (dis_vx_permute_misc(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x268: case 0x248: case 0x288: case 0x208: case 0x228: // xxlxor, xxlor, xxlnor, xxland, xxlandc
if (dis_vx_logic(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x2B2: case 0x2C0: // xsabsdp, xscpsgndp
case 0x2D2: case 0x2F2: // xsnabsdp, xsnegdp
case 0x280: case 0x2A0: // xsmaxdp, xsmindp
case 0x0F2: case 0x0D2: // xsrdpim, xsrdpip
case 0x0B4: case 0x094: // xsredp, xsrsqrtedp
case 0x0D6: case 0x0B2: // xsrdpic, xsrdpiz
case 0x092: // xsrdpi
if (dis_vxs_misc(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x08C: case 0x0AC: // xscmpudp, xscmpodp
if (dis_vx_cmp(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x080: case 0x0E0: // xsadddp, xsdivdp
case 0x084: case 0x0A4: // xsmaddadp, xsmaddmdp
case 0x0C4: case 0x0E4: // xsmsubadp, xsmsubmdp
case 0x284: case 0x2A4: // xsnmaddadp, xsnmaddmdp
case 0x2C4: case 0x2E4: // xsnmsubadp, xsnmsubmdp
case 0x0C0: case 0x0A0: // xsmuldp, xssubdp
case 0x096: case 0x0F4: // xssqrtdp, xstdivdp
case 0x0D4: // xstsqrtdp
if (dis_vxs_arith(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x180: // xvadddp
case 0x1E0: // xvdivdp
case 0x1C0: // xvmuldp
case 0x1A0: // xvsubdp
case 0x184: case 0x1A4: // xvmaddadp, xvmaddmdp
case 0x1C4: case 0x1E4: // xvmsubadp, xvmsubmdp
case 0x384: case 0x3A4: // xvnmaddadp, xvnmaddmdp
case 0x3C4: case 0x3E4: // xvnmsubadp, xvnmsubmdp
case 0x1D4: case 0x1F4: // xvtsqrtdp, xvtdivdp
case 0x196: // xvsqrtdp
if (dis_vxv_dp_arith(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x100: // xvaddsp
case 0x160: // xvdivsp
case 0x140: // xvmulsp
case 0x120: // xvsubsp
case 0x104: case 0x124: // xvmaddasp, xvmaddmsp
case 0x144: case 0x164: // xvmsubasp, xvmsubmsp
case 0x304: case 0x324: // xvnmaddasp, xvnmaddmsp
case 0x344: case 0x364: // xvnmsubasp, xvnmsubmsp
case 0x154: case 0x174: // xvtsqrtsp, xvtdivsp
case 0x116: // xvsqrtsp
if (dis_vxv_sp_arith(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x2D0: case 0x3d0: // xscvuxddp, xvcvuxddp
case 0x350: case 0x1d0: // xvcvuxdsp, xvcvuxwdp
case 0x090: // xscvdpuxws
// The above VSX conversion instructions employ some ISA 2.06
// floating point conversion instructions under the covers,
// so if allow_VX (which means "supports ISA 2.06") is not set,
// we'll skip the decode.
if (!allow_VX) goto decode_noVX;
if (dis_vx_conv(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x2B0: case 0x2F0: // xscvdpsxds, xscvsxddp
case 0x1b0: case 0x130: // xvcvdpsxws, xvcvspsxws
case 0x0b0: case 0x290: // xscvdpsxws, xscvdpuxds
case 0x212: // xscvdpsp
case 0x292: case 0x312: // xscvspdp, xvcvdpsp
case 0x390: case 0x190: // xvcvdpuxds, xvcvdpuxws
case 0x3B0: case 0x310: // xvcvdpsxds, xvcvspuxds
case 0x392: case 0x330: // xvcvspdp, xvcvspsxds
case 0x110: case 0x3f0: // xvcvspuxws, xvcvsxddp
case 0x370: case 0x1f0: // xvcvsxdsp, xvcvsxwdp
case 0x170: case 0x150: // xvcvsxwsp, xvcvuxwsp
if (dis_vx_conv(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x18C: case 0x38C: // xvcmpeqdp[.]
case 0x10C: case 0x30C: // xvcmpeqsp[.]
case 0x14C: case 0x34C: // xvcmpgesp[.]
case 0x12C: case 0x32C: // xvcmpgtsp[.]
case 0x1CC: case 0x3CC: // xvcmpgedp[.]
case 0x1AC: case 0x3AC: // xvcmpgtdp[.]
if (dis_vvec_cmp(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
case 0x134: // xvresp
case 0x1B4: // xvredp
case 0x194: case 0x114: // xvrsqrtedp, xvrsqrtesp
case 0x380: case 0x3A0: // xvmaxdp, xvmindp
case 0x300: case 0x320: // xvmaxsp, xvminsp
case 0x3C0: case 0x340: // xvcpsgndp, xvcpsgnsp
case 0x3B2: case 0x332: // xvabsdp, xvabssp
case 0x3D2: case 0x352: // xvnabsdp, xvnabssp
case 0x192: case 0x1D6: // xvrdpi, xvrdpic
case 0x1F2: case 0x1D2: // xvrdpim, xvrdpip
case 0x1B2: case 0x3F2: // xvrdpiz, xvnegdp
case 0x112: case 0x156: // xvrspi, xvrspic
case 0x172: case 0x152: // xvrspim, xvrspip
case 0x132: // xvrspiz
if (dis_vxv_misc(theInstr, vsxOpc2)) goto decode_success;
goto decode_failure;
default:
goto decode_failure;
}
break;
}
/* 64bit Integer Stores */
case 0x3E: // std, stdu
if (!mode64) goto decode_failure;
if (dis_int_store( theInstr, abiinfo )) goto decode_success;
goto decode_failure;
case 0x3F:
if (!allow_F) goto decode_noF;
/* Instrs using opc[1:5] never overlap instrs using opc[1:10],
so we can simply fall through the first switch statement */
opc2 = IFIELD(theInstr, 1, 5);
switch (opc2) {
/* Floating Point Arith Instructions */
case 0x12: case 0x14: case 0x15: // fdiv, fsub, fadd
case 0x19: // fmul
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
case 0x16: // fsqrt
if (!allow_FX) goto decode_noFX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
case 0x17: case 0x1A: // fsel, frsqrte
if (!allow_GX) goto decode_noGX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
/* Floating Point Mult-Add Instructions */
case 0x1C: case 0x1D: case 0x1E: // fmsub, fmadd, fnmsub
case 0x1F: // fnmadd
if (dis_fp_multadd(theInstr)) goto decode_success;
goto decode_failure;
case 0x18: // fre
if (!allow_GX) goto decode_noGX;
if (dis_fp_arith(theInstr)) goto decode_success;
goto decode_failure;
default:
break; // Fall through
}
opc2 = IFIELD(theInstr, 1, 10);
switch (opc2) {
/* 128-bit DFP instructions */
case 0x2: // daddq - DFP Add
case 0x202: // dsubq - DFP Subtract
case 0x22: // dmulq - DFP Mult
case 0x222: // ddivq - DFP Divide
if (!allow_DFP) goto decode_noDFP;
if (dis_dfp_arithq( theInstr ))
goto decode_success;
goto decode_failure;
case 0x162: // dxexq - DFP Extract exponent
case 0x362: // diexq - DFP Insert exponent
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_extract_insertq( theInstr ))
goto decode_success;
goto decode_failure;
case 0x82: // dcmpoq, DFP comparison ordered instruction
case 0x282: // dcmpuq, DFP comparison unordered instruction
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_compare( theInstr ) )
goto decode_success;
goto decode_failure;
case 0x102: // dctqpq - DFP convert to DFP extended
case 0x302: // drdpq - DFP round to dfp Long
case 0x122: // dctfixq - DFP convert to fixed quad
case 0x322: // dcffixq - DFP convert from fixed quad
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_fmt_convq( theInstr ))
goto decode_success;
goto decode_failure;
case 0x2A2: // dtstsfq - DFP number of significant digits
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_significant_digits(theInstr))
goto decode_success;
goto decode_failure;
case 0x142: // ddedpdq DFP Decode DPD to BCD
case 0x342: // denbcdq DFP Encode BCD to DPD
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_bcdq(theInstr))
goto decode_success;
goto decode_failure;
/* Floating Point Compare Instructions */
case 0x000: // fcmpu
case 0x020: // fcmpo
if (dis_fp_cmp(theInstr)) goto decode_success;
goto decode_failure;
case 0x080: // ftdiv
case 0x0A0: // ftsqrt
if (dis_fp_tests(theInstr)) goto decode_success;
goto decode_failure;
/* Floating Point Rounding/Conversion Instructions */
case 0x00C: // frsp
case 0x00E: // fctiw
case 0x00F: // fctiwz
case 0x32E: // fctid
case 0x32F: // fctidz
case 0x34E: // fcfid
if (dis_fp_round(theInstr)) goto decode_success;
goto decode_failure;
case 0x3CE: case 0x3AE: case 0x3AF: // fcfidu, fctidu[z] (implemented as native insns)
case 0x08F: case 0x08E: // fctiwu[z] (implemented as native insns)
if (!allow_VX) goto decode_noVX;
if (dis_fp_round(theInstr)) goto decode_success;
goto decode_failure;
/* Power6 rounding stuff */
case 0x1E8: // frim
case 0x1C8: // frip
case 0x188: // frin
case 0x1A8: // friz
/* A hack to check for P6 capability . . . */
if ((allow_F && allow_V && allow_FX && allow_GX) &&
(dis_fp_round(theInstr)))
goto decode_success;
goto decode_failure;
/* Floating Point Move Instructions */
case 0x008: // fcpsgn
case 0x028: // fneg
case 0x048: // fmr
case 0x088: // fnabs
case 0x108: // fabs
if (dis_fp_move( theInstr )) goto decode_success;
goto decode_failure;
/* Floating Point Status/Control Register Instructions */
case 0x026: // mtfsb1
case 0x040: // mcrfs
case 0x046: // mtfsb0
case 0x086: // mtfsfi
case 0x247: // mffs
case 0x2C7: // mtfsf
// Some of the above instructions need to know more about the
// ISA level supported by the host.
if (dis_fp_scr( theInstr, allow_GX )) goto decode_success;
goto decode_failure;
default:
break; // Fall through...
}
opc2 = ifieldOPClo9( theInstr );
switch (opc2) {
case 0x42: // dscli, DFP shift left
case 0x62: // dscri, DFP shift right
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_shiftq( theInstr ))
goto decode_success;
goto decode_failure;
case 0xc2: // dtstdc, DFP test data class
case 0xe2: // dtstdg, DFP test data group
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_class_test( theInstr ))
goto decode_success;
goto decode_failure;
default:
break;
}
opc2 = ifieldOPClo8( theInstr );
switch (opc2) {
case 0x3: // dquaq - DFP Quantize Quad
case 0x23: // drrndq - DFP Reround Quad
case 0x43: // dquaiq - DFP Quantize immediate Quad
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_quantize_sig_rrndq( theInstr ))
goto decode_success;
goto decode_failure;
case 0xA2: // dtstexq - DFP Test exponent Quad
if (dis_dfp_exponent_test( theInstr ) )
goto decode_success;
goto decode_failure;
case 0x63: // drintxq - DFP Round to an integer value
case 0xE3: // drintnq - DFP Round to an integer value
if (!allow_DFP)
goto decode_failure;
if (dis_dfp_roundq( theInstr ))
goto decode_success;
goto decode_failure;
default:
goto decode_failure;
}
break;
case 0x13:
switch (opc2) {
/* Condition Register Logical Instructions */
case 0x101: case 0x081: case 0x121: // crand, crandc, creqv
case 0x0E1: case 0x021: case 0x1C1: // crnand, crnor, cror
case 0x1A1: case 0x0C1: case 0x000: // crorc, crxor, mcrf
if (dis_cond_logic( theInstr )) goto decode_success;
goto decode_failure;
/* Branch Instructions */
case 0x210: case 0x010: // bcctr, bclr
if (dis_branch(theInstr, abiinfo, &dres,
resteerOkFn, callback_opaque))
goto decode_success;
goto decode_failure;
/* Memory Synchronization Instructions */
case 0x096: // isync
if (dis_memsync( theInstr )) goto decode_success;
goto decode_failure;
default:
goto decode_failure;
}
break;
case 0x1F:
/* For arith instns, bit10 is the OE flag (overflow enable) */
opc2 = IFIELD(theInstr, 1, 9);
switch (opc2) {
/* Integer Arithmetic Instructions */
case 0x10A: case 0x00A: case 0x08A: // add, addc, adde
case 0x0EA: case 0x0CA: case 0x1EB: // addme, addze, divw
case 0x1CB: case 0x04B: case 0x00B: // divwu, mulhw, mulhwu
case 0x0EB: case 0x068: case 0x028: // mullw, neg, subf
case 0x008: case 0x088: case 0x0E8: // subfc, subfe, subfme
case 0x0C8: // subfze
if (dis_int_arith( theInstr )) goto decode_success;
goto decode_failure;
case 0x18B: // divweu (implemented as native insn)
case 0x1AB: // divwe (implemented as native insn)
if (!allow_VX) goto decode_noVX;
if (dis_int_arith( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Arithmetic */
case 0x009: case 0x049: case 0x0E9: // mulhdu, mulhd, mulld
case 0x1C9: case 0x1E9: // divdu, divd
if (!mode64) goto decode_failure;
if (dis_int_arith( theInstr )) goto decode_success;
goto decode_failure;
case 0x1A9: // divde (implemented as native insn)
case 0x189: // divdeuo (implemented as native insn)
if (!allow_VX) goto decode_noVX;
if (!mode64) goto decode_failure;
if (dis_int_arith( theInstr )) goto decode_success;
goto decode_failure;
case 0x1FC: // cmpb
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
default:
break; // Fall through...
}
/* All remaining opcodes use full 10 bits. */
opc2 = IFIELD(theInstr, 1, 10);
switch (opc2) {
/* Integer Compare Instructions */
case 0x000: case 0x020: // cmp, cmpl
if (dis_int_cmp( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Logical Instructions */
case 0x01C: case 0x03C: case 0x01A: // and, andc, cntlzw
case 0x11C: case 0x3BA: case 0x39A: // eqv, extsb, extsh
case 0x1DC: case 0x07C: case 0x1BC: // nand, nor, or
case 0x19C: case 0x13C: // orc, xor
case 0x2DF: case 0x25F: // mftgpr, mffgpr
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Logical Instructions */
case 0x3DA: case 0x03A: // extsw, cntlzd
if (!mode64) goto decode_failure;
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Parity Instructions */
case 0xba: case 0x9a: // prtyd, prtyw
if (dis_int_parity( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Shift Instructions */
case 0x018: case 0x318: case 0x338: // slw, sraw, srawi
case 0x218: // srw
if (dis_int_shift( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Shift Instructions */
case 0x01B: case 0x31A: // sld, srad
case 0x33A: case 0x33B: // sradi
case 0x21B: // srd
if (!mode64) goto decode_failure;
if (dis_int_shift( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Load Instructions */
case 0x057: case 0x077: case 0x157: // lbzx, lbzux, lhax
case 0x177: case 0x117: case 0x137: // lhaux, lhzx, lhzux
case 0x017: case 0x037: // lwzx, lwzux
if (dis_int_load( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Integer Load Instructions */
case 0x035: case 0x015: // ldux, ldx
case 0x175: case 0x155: // lwaux, lwax
if (!mode64) goto decode_failure;
if (dis_int_load( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Store Instructions */
case 0x0F7: case 0x0D7: case 0x1B7: // stbux, stbx, sthux
case 0x197: case 0x0B7: case 0x097: // sthx, stwux, stwx
if (dis_int_store( theInstr, abiinfo )) goto decode_success;
goto decode_failure;
/* 64bit Integer Store Instructions */
case 0x0B5: case 0x095: // stdux, stdx
if (!mode64) goto decode_failure;
if (dis_int_store( theInstr, abiinfo )) goto decode_success;
goto decode_failure;
/* Integer Load and Store with Byte Reverse Instructions */
case 0x316: case 0x216: case 0x396: // lhbrx, lwbrx, sthbrx
case 0x296: case 0x214: // stwbrx, ldbrx
case 0x294: // stdbrx
if (dis_int_ldst_rev( theInstr )) goto decode_success;
goto decode_failure;
/* Integer Load and Store String Instructions */
case 0x255: case 0x215: case 0x2D5: // lswi, lswx, stswi
case 0x295: { // stswx
Bool stopHere = False;
Bool ok = dis_int_ldst_str( theInstr, &stopHere );
if (!ok) goto decode_failure;
if (stopHere) {
putGST( PPC_GST_CIA, mkSzImm(ty, nextInsnAddr()) );
dres.jk_StopHere = Ijk_Boring;
dres.whatNext = Dis_StopHere;
}
goto decode_success;
}
/* Memory Synchronization Instructions */
case 0x356: case 0x014: case 0x096: // eieio, lwarx, stwcx.
case 0x256: // sync
if (dis_memsync( theInstr )) goto decode_success;
goto decode_failure;
/* 64bit Memory Synchronization Instructions */
case 0x054: case 0x0D6: // ldarx, stdcx.
if (!mode64) goto decode_failure;
if (dis_memsync( theInstr )) goto decode_success;
goto decode_failure;
/* Processor Control Instructions */
case 0x200: case 0x013: case 0x153: // mcrxr, mfcr, mfspr
case 0x173: case 0x090: case 0x1D3: // mftb, mtcrf, mtspr
if (dis_proc_ctl( abiinfo, theInstr )) goto decode_success;
goto decode_failure;
/* Cache Management Instructions */
case 0x2F6: case 0x056: case 0x036: // dcba, dcbf, dcbst
case 0x116: case 0x0F6: case 0x3F6: // dcbt, dcbtst, dcbz
case 0x3D6: // icbi
if (dis_cache_manage( theInstr, &dres, archinfo ))
goto decode_success;
goto decode_failure;
//zz /* External Control Instructions */
//zz case 0x136: case 0x1B6: // eciwx, ecowx
//zz DIP("external control op => not implemented\n");
//zz goto decode_failure;
/* Trap Instructions */
case 0x004: case 0x044: // tw, td
if (dis_trap(theInstr, &dres)) goto decode_success;
goto decode_failure;
/* Floating Point Load Instructions */
case 0x217: case 0x237: case 0x257: // lfsx, lfsux, lfdx
case 0x277: // lfdux
if (!allow_F) goto decode_noF;
if (dis_fp_load( theInstr )) goto decode_success;
goto decode_failure;
/* Floating Point Store Instructions */
case 0x297: case 0x2B7: case 0x2D7: // stfs, stfsu, stfd
case 0x2F7: // stfdu, stfiwx
if (!allow_F) goto decode_noF;
if (dis_fp_store( theInstr )) goto decode_success;
goto decode_failure;
case 0x3D7: // stfiwx
if (!allow_F) goto decode_noF;
if (!allow_GX) goto decode_noGX;
if (dis_fp_store( theInstr )) goto decode_success;
goto decode_failure;
/* Floating Point Double Pair Indexed Instructions */
case 0x317: // lfdpx (Power6)
case 0x397: // stfdpx (Power6)
if (!allow_F) goto decode_noF;
if (dis_fp_pair(theInstr)) goto decode_success;
goto decode_failure;
case 0x357: // lfiwax
if (!allow_F) goto decode_noF;
if (dis_fp_load( theInstr )) goto decode_success;
goto decode_failure;
case 0x377: // lfiwzx
if (!allow_F) goto decode_noF;
if (dis_fp_load( theInstr )) goto decode_success;
goto decode_failure;
/* AltiVec instructions */
/* AV Cache Control - Data streams */
case 0x156: case 0x176: case 0x336: // dst, dstst, dss
if (!allow_V) goto decode_noV;
if (dis_av_datastream( theInstr )) goto decode_success;
goto decode_failure;
/* AV Load */
case 0x006: case 0x026: // lvsl, lvsr
case 0x007: case 0x027: case 0x047: // lvebx, lvehx, lvewx
case 0x067: case 0x167: // lvx, lvxl
if (!allow_V) goto decode_noV;
if (dis_av_load( abiinfo, theInstr )) goto decode_success;
goto decode_failure;
/* AV Store */
case 0x087: case 0x0A7: case 0x0C7: // stvebx, stvehx, stvewx
case 0x0E7: case 0x1E7: // stvx, stvxl
if (!allow_V) goto decode_noV;
if (dis_av_store( theInstr )) goto decode_success;
goto decode_failure;
/* VSX Load */
case 0x24C: // lxsdx
case 0x34C: // lxvd2x
case 0x14C: // lxvdsx
case 0x30C: // lxvw4x
// All of these VSX load instructions use some VMX facilities, so
// if allow_V is not set, we'll skip trying to decode.
if (!allow_V) goto decode_noV;
if (dis_vx_load( theInstr )) goto decode_success;
goto decode_failure;
/* VSX Store */
case 0x2CC: // stxsdx
case 0x3CC: // stxvd2x
case 0x38C: // stxvw4x
// All of these VSX store instructions use some VMX facilities, so
// if allow_V is not set, we'll skip trying to decode.
if (!allow_V) goto decode_noV;
if (dis_vx_store( theInstr )) goto decode_success;
goto decode_failure;
/* Miscellaneous ISA 2.06 instructions */
case 0x1FA: // popcntd
case 0x17A: // popcntw
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
case 0x0FC: // bpermd
if (dis_int_logic( theInstr )) goto decode_success;
goto decode_failure;
default:
/* Deal with some other cases that we would otherwise have
punted on. */
/* --- ISEL (PowerISA_V2.05.pdf, p74) --- */
/* only decode this insn when reserved bit 0 (31 in IBM's
notation) is zero */
if (IFIELD(theInstr, 0, 6) == (15<<1)) {
UInt rT = ifieldRegDS( theInstr );
UInt rA = ifieldRegA( theInstr );
UInt rB = ifieldRegB( theInstr );
UInt bi = ifieldRegC( theInstr );
putIReg(
rT,
IRExpr_Mux0X( unop(Iop_32to8,getCRbit( bi )),
getIReg(rB),
rA == 0 ? (mode64 ? mkU64(0) : mkU32(0))
: getIReg(rA) )
);
DIP("isel r%u,r%u,r%u,crb%u\n", rT,rA,rB,bi);
goto decode_success;
}
goto decode_failure;
}
break;
case 0x04:
/* AltiVec instructions */
opc2 = IFIELD(theInstr, 0, 6);
switch (opc2) {
/* AV Mult-Add, Mult-Sum */
case 0x20: case 0x21: case 0x22: // vmhaddshs, vmhraddshs, vmladduhm
case 0x24: case 0x25: case 0x26: // vmsumubm, vmsummbm, vmsumuhm
case 0x27: case 0x28: case 0x29: // vmsumuhs, vmsumshm, vmsumshs
if (!allow_V) goto decode_noV;
if (dis_av_multarith( theInstr )) goto decode_success;
goto decode_failure;
/* AV Permutations */
case 0x2A: // vsel
case 0x2B: // vperm
case 0x2C: // vsldoi
if (!allow_V) goto decode_noV;
if (dis_av_permute( theInstr )) goto decode_success;
goto decode_failure;
/* AV Floating Point Mult-Add/Sub */
case 0x2E: case 0x2F: // vmaddfp, vnmsubfp
if (!allow_V) goto decode_noV;
if (dis_av_fp_arith( theInstr )) goto decode_success;
goto decode_failure;
default:
break; // Fall through...
}
opc2 = IFIELD(theInstr, 0, 11);
switch (opc2) {
/* AV Arithmetic */
case 0x180: // vaddcuw
case 0x000: case 0x040: case 0x080: // vaddubm, vadduhm, vadduwm
case 0x200: case 0x240: case 0x280: // vaddubs, vadduhs, vadduws
case 0x300: case 0x340: case 0x380: // vaddsbs, vaddshs, vaddsws
case 0x580: // vsubcuw
case 0x400: case 0x440: case 0x480: // vsububm, vsubuhm, vsubuwm
case 0x600: case 0x640: case 0x680: // vsububs, vsubuhs, vsubuws
case 0x700: case 0x740: case 0x780: // vsubsbs, vsubshs, vsubsws
case 0x402: case 0x442: case 0x482: // vavgub, vavguh, vavguw
case 0x502: case 0x542: case 0x582: // vavgsb, vavgsh, vavgsw
case 0x002: case 0x042: case 0x082: // vmaxub, vmaxuh, vmaxuw
case 0x102: case 0x142: case 0x182: // vmaxsb, vmaxsh, vmaxsw
case 0x202: case 0x242: case 0x282: // vminub, vminuh, vminuw
case 0x302: case 0x342: case 0x382: // vminsb, vminsh, vminsw
case 0x008: case 0x048: // vmuloub, vmulouh
case 0x108: case 0x148: // vmulosb, vmulosh
case 0x208: case 0x248: // vmuleub, vmuleuh
case 0x308: case 0x348: // vmulesb, vmulesh
case 0x608: case 0x708: case 0x648: // vsum4ubs, vsum4sbs, vsum4shs
case 0x688: case 0x788: // vsum2sws, vsumsws
if (!allow_V) goto decode_noV;
if (dis_av_arith( theInstr )) goto decode_success;
goto decode_failure;
/* AV Rotate, Shift */
case 0x004: case 0x044: case 0x084: // vrlb, vrlh, vrlw
case 0x104: case 0x144: case 0x184: // vslb, vslh, vslw
case 0x204: case 0x244: case 0x284: // vsrb, vsrh, vsrw
case 0x304: case 0x344: case 0x384: // vsrab, vsrah, vsraw
case 0x1C4: case 0x2C4: // vsl, vsr
case 0x40C: case 0x44C: // vslo, vsro
if (!allow_V) goto decode_noV;
if (dis_av_shift( theInstr )) goto decode_success;
goto decode_failure;
/* AV Logic */
case 0x404: case 0x444: case 0x484: // vand, vandc, vor
case 0x4C4: case 0x504: // vxor, vnor
if (!allow_V) goto decode_noV;
if (dis_av_logic( theInstr )) goto decode_success;
goto decode_failure;
/* AV Processor Control */
case 0x604: case 0x644: // mfvscr, mtvscr
if (!allow_V) goto decode_noV;
if (dis_av_procctl( theInstr )) goto decode_success;
goto decode_failure;
/* AV Floating Point Arithmetic */
case 0x00A: case 0x04A: // vaddfp, vsubfp
case 0x10A: case 0x14A: case 0x18A: // vrefp, vrsqrtefp, vexptefp
case 0x1CA: // vlogefp
case 0x40A: case 0x44A: // vmaxfp, vminfp
if (!allow_V) goto decode_noV;
if (dis_av_fp_arith( theInstr )) goto decode_success;
goto decode_failure;
/* AV Floating Point Round/Convert */
case 0x20A: case 0x24A: case 0x28A: // vrfin, vrfiz, vrfip
case 0x2CA: // vrfim
case 0x30A: case 0x34A: case 0x38A: // vcfux, vcfsx, vctuxs
case 0x3CA: // vctsxs
if (!allow_V) goto decode_noV;
if (dis_av_fp_convert( theInstr )) goto decode_success;
goto decode_failure;
/* AV Merge, Splat */
case 0x00C: case 0x04C: case 0x08C: // vmrghb, vmrghh, vmrghw
case 0x10C: case 0x14C: case 0x18C: // vmrglb, vmrglh, vmrglw
case 0x20C: case 0x24C: case 0x28C: // vspltb, vsplth, vspltw
case 0x30C: case 0x34C: case 0x38C: // vspltisb, vspltish, vspltisw
if (!allow_V) goto decode_noV;
if (dis_av_permute( theInstr )) goto decode_success;
goto decode_failure;
/* AV Pack, Unpack */
case 0x00E: case 0x04E: case 0x08E: // vpkuhum, vpkuwum, vpkuhus
case 0x0CE: // vpkuwus
case 0x10E: case 0x14E: case 0x18E: // vpkshus, vpkswus, vpkshss
case 0x1CE: // vpkswss
case 0x20E: case 0x24E: case 0x28E: // vupkhsb, vupkhsh, vupklsb
case 0x2CE: // vupklsh
case 0x30E: case 0x34E: case 0x3CE: // vpkpx, vupkhpx, vupklpx
if (!allow_V) goto decode_noV;
if (dis_av_pack( theInstr )) goto decode_success;
goto decode_failure;
default:
break; // Fall through...
}
opc2 = IFIELD(theInstr, 0, 10);
switch (opc2) {
/* AV Compare */
case 0x006: case 0x046: case 0x086: // vcmpequb, vcmpequh, vcmpequw
case 0x206: case 0x246: case 0x286: // vcmpgtub, vcmpgtuh, vcmpgtuw
case 0x306: case 0x346: case 0x386: // vcmpgtsb, vcmpgtsh, vcmpgtsw
if (!allow_V) goto decode_noV;
if (dis_av_cmp( theInstr )) goto decode_success;
goto decode_failure;
/* AV Floating Point Compare */
case 0x0C6: case 0x1C6: case 0x2C6: // vcmpeqfp, vcmpgefp, vcmpgtfp
case 0x3C6: // vcmpbfp
if (!allow_V) goto decode_noV;
if (dis_av_fp_cmp( theInstr )) goto decode_success;
goto decode_failure;
default:
goto decode_failure;
}
break;
default:
goto decode_failure;
decode_noF:
vassert(!allow_F);
vex_printf("disInstr(ppc): declined to decode an FP insn.\n");
goto decode_failure;
decode_noV:
vassert(!allow_V);
vex_printf("disInstr(ppc): declined to decode an AltiVec insn.\n");
goto decode_failure;
decode_noVX:
vassert(!allow_VX);
vex_printf("disInstr(ppc): declined to decode a Power ISA 2.06 insn.\n");
goto decode_failure;
decode_noFX:
vassert(!allow_FX);
vex_printf("disInstr(ppc): "
"declined to decode a GeneralPurpose-Optional insn.\n");
goto decode_failure;
decode_noGX:
vassert(!allow_GX);
vex_printf("disInstr(ppc): "
"declined to decode a Graphics-Optional insn.\n");
goto decode_failure;
decode_noDFP:
vassert(!allow_DFP);
vex_printf("disInstr(ppc): "
"declined to decode a Decimal Floating Point insn.\n");
goto decode_failure;
decode_failure:
/* All decode failures end up here. */
opc2 = (theInstr) & 0x7FF;
vex_printf("disInstr(ppc): unhandled instruction: "
"0x%x\n", theInstr);
vex_printf(" primary %d(0x%x), secondary %u(0x%x)\n",
opc1, opc1, opc2, opc2);
/* Tell the dispatcher that this insn cannot be decoded, and so has
not been executed, and (is currently) the next to be executed.
CIA should be up-to-date since it made so at the start of each
insn, but nevertheless be paranoid and update it again right
now. */
putGST( PPC_GST_CIA, mkSzImm(ty, guest_CIA_curr_instr) );
dres.whatNext = Dis_StopHere;
dres.jk_StopHere = Ijk_NoDecode;
dres.len = 0;
return dres;
} /* switch (opc) for the main (primary) opcode switch. */
decode_success:
/* All decode successes end up here. */
switch (dres.whatNext) {
case Dis_Continue:
putGST( PPC_GST_CIA, mkSzImm(ty, guest_CIA_curr_instr + 4));
break;
case Dis_ResteerU:
case Dis_ResteerC:
putGST( PPC_GST_CIA, mkSzImm(ty, dres.continueAt));
break;
case Dis_StopHere:
break;
default:
vassert(0);
}
DIP("\n");
if (dres.len == 0) {
dres.len = 4;
} else {
vassert(dres.len == 20);
}
return dres;
}
#undef DIP
#undef DIS
/*------------------------------------------------------------*/
/*--- Top-level fn ---*/
/*------------------------------------------------------------*/
/* Disassemble a single instruction into IR. The instruction
is located in host memory at &guest_code[delta]. */
DisResult disInstr_PPC ( IRSB* irsb_IN,
Bool (*resteerOkFn) ( void*, Addr64 ),
Bool resteerCisOk,
void* callback_opaque,
UChar* guest_code_IN,
Long delta,
Addr64 guest_IP,
VexArch guest_arch,
VexArchInfo* archinfo,
VexAbiInfo* abiinfo,
Bool host_bigendian_IN )
{
IRType ty;
DisResult dres;
UInt mask32, mask64;
UInt hwcaps_guest = archinfo->hwcaps;
vassert(guest_arch == VexArchPPC32 || guest_arch == VexArchPPC64);
/* global -- ick */
mode64 = guest_arch == VexArchPPC64;
ty = mode64 ? Ity_I64 : Ity_I32;
/* do some sanity checks */
mask32 = VEX_HWCAPS_PPC32_F | VEX_HWCAPS_PPC32_V
| VEX_HWCAPS_PPC32_FX | VEX_HWCAPS_PPC32_GX | VEX_HWCAPS_PPC32_VX
| VEX_HWCAPS_PPC32_DFP;
mask64 = VEX_HWCAPS_PPC64_V | VEX_HWCAPS_PPC64_FX
| VEX_HWCAPS_PPC64_GX | VEX_HWCAPS_PPC64_VX | VEX_HWCAPS_PPC64_DFP;
if (mode64) {
vassert((hwcaps_guest & mask32) == 0);
} else {
vassert((hwcaps_guest & mask64) == 0);
}
/* Set globals (see top of this file) */
guest_code = guest_code_IN;
irsb = irsb_IN;
host_is_bigendian = host_bigendian_IN;
guest_CIA_curr_instr = mkSzAddr(ty, guest_IP);
guest_CIA_bbstart = mkSzAddr(ty, guest_IP - delta);
dres = disInstr_PPC_WRK ( resteerOkFn, resteerCisOk, callback_opaque,
delta, archinfo, abiinfo );
return dres;
}
/*------------------------------------------------------------*/
/*--- Unused stuff ---*/
/*------------------------------------------------------------*/
///* A potentially more memcheck-friendly implementation of Clz32, with
// the boundary case Clz32(0) = 32, which is what ppc requires. */
//
//static IRExpr* /* :: Ity_I32 */ verbose_Clz32 ( IRTemp arg )
//{
// /* Welcome ... to SSA R Us. */
// IRTemp n1 = newTemp(Ity_I32);
// IRTemp n2 = newTemp(Ity_I32);
// IRTemp n3 = newTemp(Ity_I32);
// IRTemp n4 = newTemp(Ity_I32);
// IRTemp n5 = newTemp(Ity_I32);
// IRTemp n6 = newTemp(Ity_I32);
// IRTemp n7 = newTemp(Ity_I32);
// IRTemp n8 = newTemp(Ity_I32);
// IRTemp n9 = newTemp(Ity_I32);
// IRTemp n10 = newTemp(Ity_I32);
// IRTemp n11 = newTemp(Ity_I32);
// IRTemp n12 = newTemp(Ity_I32);
//
// /* First, propagate the most significant 1-bit into all lower
// positions in the word. */
// /* unsigned int clz ( unsigned int n )
// {
// n |= (n >> 1);
// n |= (n >> 2);
// n |= (n >> 4);
// n |= (n >> 8);
// n |= (n >> 16);
// return bitcount(~n);
// }
// */
// assign(n1, mkexpr(arg));
// assign(n2, binop(Iop_Or32, mkexpr(n1), binop(Iop_Shr32, mkexpr(n1), mkU8(1))));
// assign(n3, binop(Iop_Or32, mkexpr(n2), binop(Iop_Shr32, mkexpr(n2), mkU8(2))));
// assign(n4, binop(Iop_Or32, mkexpr(n3), binop(Iop_Shr32, mkexpr(n3), mkU8(4))));
// assign(n5, binop(Iop_Or32, mkexpr(n4), binop(Iop_Shr32, mkexpr(n4), mkU8(8))));
// assign(n6, binop(Iop_Or32, mkexpr(n5), binop(Iop_Shr32, mkexpr(n5), mkU8(16))));
// /* This gives a word of the form 0---01---1. Now invert it, giving
// a word of the form 1---10---0, then do a population-count idiom
// (to count the 1s, which is the number of leading zeroes, or 32
// if the original word was 0. */
// assign(n7, unop(Iop_Not32, mkexpr(n6)));
//
// /* unsigned int bitcount ( unsigned int n )
// {
// n = n - ((n >> 1) & 0x55555555);
// n = (n & 0x33333333) + ((n >> 2) & 0x33333333);
// n = (n + (n >> 4)) & 0x0F0F0F0F;
// n = n + (n >> 8);
// n = (n + (n >> 16)) & 0x3F;
// return n;
// }
// */
// assign(n8,
// binop(Iop_Sub32,
// mkexpr(n7),
// binop(Iop_And32,
// binop(Iop_Shr32, mkexpr(n7), mkU8(1)),
// mkU32(0x55555555))));
// assign(n9,
// binop(Iop_Add32,
// binop(Iop_And32, mkexpr(n8), mkU32(0x33333333)),
// binop(Iop_And32,
// binop(Iop_Shr32, mkexpr(n8), mkU8(2)),
// mkU32(0x33333333))));
// assign(n10,
// binop(Iop_And32,
// binop(Iop_Add32,
// mkexpr(n9),
// binop(Iop_Shr32, mkexpr(n9), mkU8(4))),
// mkU32(0x0F0F0F0F)));
// assign(n11,
// binop(Iop_Add32,
// mkexpr(n10),
// binop(Iop_Shr32, mkexpr(n10), mkU8(8))));
// assign(n12,
// binop(Iop_Add32,
// mkexpr(n11),
// binop(Iop_Shr32, mkexpr(n11), mkU8(16))));
// return
// binop(Iop_And32, mkexpr(n12), mkU32(0x3F));
//}
/*--------------------------------------------------------------------*/
/*--- end guest_ppc_toIR.c ---*/
/*--------------------------------------------------------------------*/