1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240 |
- /*
- * Copyright (C) 2013 The Android Open Source Project
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- * * Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
- * distribution.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
- * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
- * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
- * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
- * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
- * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
- * SUCH DAMAGE.
- */
- #define LOG_TAG "ArmToArm64Assembler"
- #include <stdio.h>
- #include <stdlib.h>
- #include <string.h>
- #include <cutils/properties.h>
- #include <log/log.h>
- #include <private/pixelflinger/ggl_context.h>
- #include "codeflinger/Arm64Assembler.h"
- #include "codeflinger/Arm64Disassembler.h"
- #include "codeflinger/CodeCache.h"
- /*
- ** --------------------------------------------
- ** Support for Arm64 in GGLAssembler JIT
- ** --------------------------------------------
- **
- ** Approach
- ** - GGLAssembler and associated files are largely un-changed.
- ** - A translator class maps ArmAssemblerInterface calls to
- ** generate Arm64 instructions.
- **
- ** ----------------------
- ** ArmToArm64Assembler
- ** ----------------------
- **
- ** - Subclassed from ArmAssemblerInterface
- **
- ** - Translates each ArmAssemblerInterface call to generate
- ** one or more Arm64 instructions as necessary.
- **
- ** - Does not implement ArmAssemblerInterface portions unused by GGLAssembler
- ** It calls NOT_IMPLEMENTED() for such cases, which in turn logs
- ** a fatal message.
- **
- ** - Uses A64_.. series of functions to generate instruction machine code
- ** for Arm64 instructions. These functions also log the instruction
- ** to LOG, if ARM64_ASM_DEBUG define is set to 1
- **
- ** - Dumps machine code and eqvt assembly if "debug.pf.disasm" option is set
- ** It uses arm64_disassemble to perform disassembly
- **
- ** - Uses register 13 (SP in ARM), 15 (PC in ARM), 16, 17 for storing
- ** intermediate results. GGLAssembler does not use SP and PC as these
- ** registers are marked as reserved. The temporary registers are not
- ** saved/restored on stack as these are caller-saved registers in Arm64
- **
- ** - Uses CSEL instruction to support conditional execution. The result is
- ** stored in a temporary register and then copied to the target register
- ** if the condition is true.
- **
- ** - In the case of conditional data transfer instructions, conditional
- ** branch is used to skip over instruction, if the condition is false
- **
- ** - Wherever possible, immediate values are transferred to temporary
- ** register prior to processing. This simplifies overall implementation
- ** as instructions requiring immediate values are converted to
- ** move immediate instructions followed by register-register instruction.
- **
- ** --------------------------------------------
- ** ArmToArm64Assembler unit test bench
- ** --------------------------------------------
- **
- ** - Tests ArmToArm64Assembler interface for all the possible
- ** ways in which GGLAssembler uses ArmAssemblerInterface interface.
- **
- ** - Uses test jacket (written in assembly) to set the registers,
- ** condition flags prior to calling generated instruction. It also
- ** copies registers and flags at the end of execution. Caller then
- ** checks if generated code performed correct operation based on
- ** output registers and flags.
- **
- ** - Broadly contains three type of tests, (i) data operation tests
- ** (ii) data transfer tests and (iii) LDM/STM tests.
- **
- ** ----------------------
- ** Arm64 disassembler
- ** ----------------------
- ** - This disassembler disassembles only those machine codes which can be
- ** generated by ArmToArm64Assembler. It has a unit testbench which
- ** tests all the instructions supported by the disassembler.
- **
- ** ------------------------------------------------------------------
- ** ARMAssembler/ARMAssemblerInterface/ARMAssemblerProxy changes
- ** ------------------------------------------------------------------
- **
- ** - In existing code, addresses were being handled as 32 bit values at
- ** certain places.
- **
- ** - Added a new set of functions for address load/store/manipulation.
- ** These are ADDR_LDR, ADDR_STR, ADDR_ADD, ADDR_SUB and they map to
- ** default 32 bit implementations in ARMAssemblerInterface.
- **
- ** - ArmToArm64Assembler maps these functions to appropriate 64 bit
- ** functions.
- **
- ** ----------------------
- ** GGLAssembler changes
- ** ----------------------
- ** - Since ArmToArm64Assembler can generate 4 Arm64 instructions for
- ** each call in worst case, the memory required is set to 4 times
- ** ARM memory
- **
- ** - Address load/store/manipulation were changed to use new functions
- ** added in the ARMAssemblerInterface.
- **
- */
- #define NOT_IMPLEMENTED() LOG_FATAL("Arm instruction %s not yet implemented\n", __func__)
- #define ARM64_ASM_DEBUG 0
- #if ARM64_ASM_DEBUG
- #define LOG_INSTR(...) ALOGD("\t" __VA_ARGS__)
- #define LOG_LABEL(...) ALOGD(__VA_ARGS__)
- #else
- #define LOG_INSTR(...) ((void)0)
- #define LOG_LABEL(...) ((void)0)
- #endif
- namespace android {
- static __unused const char* shift_codes[] =
- {
- "LSL", "LSR", "ASR", "ROR"
- };
- static __unused const char *cc_codes[] =
- {
- "EQ", "NE", "CS", "CC", "MI",
- "PL", "VS", "VC", "HI", "LS",
- "GE", "LT", "GT", "LE", "AL", "NV"
- };
- ArmToArm64Assembler::ArmToArm64Assembler(const sp<Assembly>& assembly)
- : ARMAssemblerInterface(),
- mAssembly(assembly)
- {
- mBase = mPC = (uint32_t *)assembly->base();
- mDuration = ggl_system_time();
- mZeroReg = 13;
- mTmpReg1 = 15;
- mTmpReg2 = 16;
- mTmpReg3 = 17;
- }
- ArmToArm64Assembler::ArmToArm64Assembler(void *base)
- : ARMAssemblerInterface(), mAssembly(NULL)
- {
- mBase = mPC = (uint32_t *)base;
- mDuration = ggl_system_time();
- // Regs 13, 15, 16, 17 are used as temporary registers
- mZeroReg = 13;
- mTmpReg1 = 15;
- mTmpReg2 = 16;
- mTmpReg3 = 17;
- }
- ArmToArm64Assembler::~ArmToArm64Assembler()
- {
- }
- uint32_t* ArmToArm64Assembler::pc() const
- {
- return mPC;
- }
- uint32_t* ArmToArm64Assembler::base() const
- {
- return mBase;
- }
- void ArmToArm64Assembler::reset()
- {
- if(mAssembly == NULL)
- mPC = mBase;
- else
- mBase = mPC = (uint32_t *)mAssembly->base();
- mBranchTargets.clear();
- mLabels.clear();
- mLabelsInverseMapping.clear();
- mComments.clear();
- #if ARM64_ASM_DEBUG
- ALOGI("RESET\n");
- #endif
- }
- int ArmToArm64Assembler::getCodegenArch()
- {
- return CODEGEN_ARCH_ARM64;
- }
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::disassemble(const char* name)
- {
- if(name)
- {
- printf("%s:\n", name);
- }
- size_t count = pc()-base();
- uint32_t* i = base();
- while (count--)
- {
- ssize_t label = mLabelsInverseMapping.indexOfKey(i);
- if (label >= 0)
- {
- printf("%s:\n", mLabelsInverseMapping.valueAt(label));
- }
- ssize_t comment = mComments.indexOfKey(i);
- if (comment >= 0)
- {
- printf("; %s\n", mComments.valueAt(comment));
- }
- printf("%p: %08x ", i, uint32_t(i[0]));
- {
- char instr[256];
- ::arm64_disassemble(*i, instr);
- printf("%s\n", instr);
- }
- i++;
- }
- }
- void ArmToArm64Assembler::comment(const char* string)
- {
- mComments.add(mPC, string);
- LOG_INSTR("//%s\n", string);
- }
- void ArmToArm64Assembler::label(const char* theLabel)
- {
- mLabels.add(theLabel, mPC);
- mLabelsInverseMapping.add(mPC, theLabel);
- LOG_LABEL("%s:\n", theLabel);
- }
- void ArmToArm64Assembler::B(int cc, const char* label)
- {
- mBranchTargets.add(branch_target_t(label, mPC));
- LOG_INSTR("B%s %s\n", cc_codes[cc], label );
- *mPC++ = (0x54 << 24) | cc;
- }
- void ArmToArm64Assembler::BL(int /*cc*/, const char* /*label*/)
- {
- NOT_IMPLEMENTED(); //Not Required
- }
- // ----------------------------------------------------------------------------
- //Prolog/Epilog & Generate...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::prolog()
- {
- // write prolog code
- mPrologPC = mPC;
- *mPC++ = A64_MOVZ_X(mZeroReg,0,0);
- }
- void ArmToArm64Assembler::epilog(uint32_t /*touched*/)
- {
- // write epilog code
- static const int XLR = 30;
- *mPC++ = A64_RET(XLR);
- }
- int ArmToArm64Assembler::generate(const char* name)
- {
- // fixup all the branches
- size_t count = mBranchTargets.size();
- while (count--)
- {
- const branch_target_t& bt = mBranchTargets[count];
- uint32_t* target_pc = mLabels.valueFor(bt.label);
- LOG_ALWAYS_FATAL_IF(!target_pc,
- "error resolving branch targets, target_pc is null");
- int32_t offset = int32_t(target_pc - bt.pc);
- *bt.pc |= (offset & 0x7FFFF) << 5;
- }
- if(mAssembly != NULL)
- mAssembly->resize( int(pc()-base())*4 );
- // the instruction cache is flushed by CodeCache
- const int64_t duration = ggl_system_time() - mDuration;
- const char * const format = "generated %s (%d ins) at [%p:%p] in %ld ns\n";
- ALOGI(format, name, int(pc()-base()), base(), pc(), duration);
- char value[PROPERTY_VALUE_MAX];
- property_get("debug.pf.disasm", value, "0");
- if (atoi(value) != 0)
- {
- printf(format, name, int(pc()-base()), base(), pc(), duration);
- disassemble(name);
- }
- return OK;
- }
- uint32_t* ArmToArm64Assembler::pcForLabel(const char* label)
- {
- return mLabels.valueFor(label);
- }
- // ----------------------------------------------------------------------------
- // Data Processing...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::dataProcessingCommon(int opcode,
- int s, int Rd, int Rn, uint32_t Op2)
- {
- if(opcode != opSUB && s == 1)
- {
- NOT_IMPLEMENTED(); //Not required
- return;
- }
- if(opcode != opSUB && opcode != opADD && opcode != opAND &&
- opcode != opORR && opcode != opMVN)
- {
- NOT_IMPLEMENTED(); //Not required
- return;
- }
- if(Op2 == OPERAND_REG_IMM && mAddrMode.reg_imm_shift > 31)
- {
- NOT_IMPLEMENTED();
- return;
- }
- //Store immediate in temporary register and convert
- //immediate operation into register operation
- if(Op2 == OPERAND_IMM)
- {
- int imm = mAddrMode.immediate;
- *mPC++ = A64_MOVZ_W(mTmpReg2, imm & 0x0000FFFF, 0);
- *mPC++ = A64_MOVK_W(mTmpReg2, (imm >> 16) & 0x0000FFFF, 16);
- Op2 = mTmpReg2;
- }
- {
- uint32_t shift;
- uint32_t amount;
- uint32_t Rm;
- if(Op2 == OPERAND_REG_IMM)
- {
- shift = mAddrMode.reg_imm_type;
- amount = mAddrMode.reg_imm_shift;
- Rm = mAddrMode.reg_imm_Rm;
- }
- else if(Op2 < OPERAND_REG)
- {
- shift = 0;
- amount = 0;
- Rm = Op2;
- }
- else
- {
- NOT_IMPLEMENTED(); //Not required
- return;
- }
- switch(opcode)
- {
- case opADD: *mPC++ = A64_ADD_W(Rd, Rn, Rm, shift, amount); break;
- case opAND: *mPC++ = A64_AND_W(Rd, Rn, Rm, shift, amount); break;
- case opORR: *mPC++ = A64_ORR_W(Rd, Rn, Rm, shift, amount); break;
- case opMVN: *mPC++ = A64_ORN_W(Rd, Rn, Rm, shift, amount); break;
- case opSUB: *mPC++ = A64_SUB_W(Rd, Rn, Rm, shift, amount, s);break;
- };
- }
- }
- void ArmToArm64Assembler::dataProcessing(int opcode, int cc,
- int s, int Rd, int Rn, uint32_t Op2)
- {
- uint32_t Wd;
- if(cc != AL)
- Wd = mTmpReg1;
- else
- Wd = Rd;
- if(opcode == opADD || opcode == opAND || opcode == opORR ||opcode == opSUB)
- {
- dataProcessingCommon(opcode, s, Wd, Rn, Op2);
- }
- else if(opcode == opCMP)
- {
- dataProcessingCommon(opSUB, 1, mTmpReg3, Rn, Op2);
- }
- else if(opcode == opRSB)
- {
- dataProcessingCommon(opSUB, s, Wd, Rn, Op2);
- dataProcessingCommon(opSUB, s, Wd, mZeroReg, Wd);
- }
- else if(opcode == opMOV)
- {
- dataProcessingCommon(opORR, 0, Wd, mZeroReg, Op2);
- if(s == 1)
- {
- dataProcessingCommon(opSUB, 1, mTmpReg3, Wd, mZeroReg);
- }
- }
- else if(opcode == opMVN)
- {
- dataProcessingCommon(opMVN, s, Wd, mZeroReg, Op2);
- }
- else if(opcode == opBIC)
- {
- dataProcessingCommon(opMVN, s, mTmpReg3, mZeroReg, Op2);
- dataProcessingCommon(opAND, s, Wd, Rn, mTmpReg3);
- }
- else
- {
- NOT_IMPLEMENTED();
- return;
- }
- if(cc != AL)
- {
- *mPC++ = A64_CSEL_W(Rd, mTmpReg1, Rd, cc);
- }
- }
- // ----------------------------------------------------------------------------
- // Address Processing...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::ADDR_ADD(int cc,
- int s, int Rd, int Rn, uint32_t Op2)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if(s != 0) { NOT_IMPLEMENTED(); return;} //Not required
- if(Op2 == OPERAND_REG_IMM && mAddrMode.reg_imm_type == LSL)
- {
- int Rm = mAddrMode.reg_imm_Rm;
- int amount = mAddrMode.reg_imm_shift;
- *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
- }
- else if(Op2 < OPERAND_REG)
- {
- int Rm = Op2;
- int amount = 0;
- *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
- }
- else if(Op2 == OPERAND_IMM)
- {
- int imm = mAddrMode.immediate;
- *mPC++ = A64_MOVZ_W(mTmpReg1, imm & 0x0000FFFF, 0);
- *mPC++ = A64_MOVK_W(mTmpReg1, (imm >> 16) & 0x0000FFFF, 16);
- int Rm = mTmpReg1;
- int amount = 0;
- *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
- }
- else
- {
- NOT_IMPLEMENTED(); //Not required
- }
- }
- void ArmToArm64Assembler::ADDR_SUB(int cc,
- int s, int Rd, int Rn, uint32_t Op2)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if(s != 0) { NOT_IMPLEMENTED(); return;} //Not required
- if(Op2 == OPERAND_REG_IMM && mAddrMode.reg_imm_type == LSR)
- {
- *mPC++ = A64_ADD_W(mTmpReg1, mZeroReg, mAddrMode.reg_imm_Rm,
- LSR, mAddrMode.reg_imm_shift);
- *mPC++ = A64_SUB_X_Wm_SXTW(Rd, Rn, mTmpReg1, 0);
- }
- else
- {
- NOT_IMPLEMENTED(); //Not required
- }
- }
- // ----------------------------------------------------------------------------
- // multiply...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::MLA(int cc, int s,int Rd, int Rm, int Rs, int Rn)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- *mPC++ = A64_MADD_W(Rd, Rm, Rs, Rn);
- if(s == 1)
- dataProcessingCommon(opSUB, 1, mTmpReg1, Rd, mZeroReg);
- }
- void ArmToArm64Assembler::MUL(int cc, int s, int Rd, int Rm, int Rs)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if(s != 0) { NOT_IMPLEMENTED(); return;} //Not required
- *mPC++ = A64_MADD_W(Rd, Rm, Rs, mZeroReg);
- }
- void ArmToArm64Assembler::UMULL(int /*cc*/, int /*s*/,
- int /*RdLo*/, int /*RdHi*/, int /*Rm*/, int /*Rs*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::UMUAL(int /*cc*/, int /*s*/,
- int /*RdLo*/, int /*RdHi*/, int /*Rm*/, int /*Rs*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::SMULL(int /*cc*/, int /*s*/,
- int /*RdLo*/, int /*RdHi*/, int /*Rm*/, int /*Rs*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::SMUAL(int /*cc*/, int /*s*/,
- int /*RdLo*/, int /*RdHi*/, int /*Rm*/, int /*Rs*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- // ----------------------------------------------------------------------------
- // branches relative to PC...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::B(int /*cc*/, uint32_t* /*pc*/){
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::BL(int /*cc*/, uint32_t* /*pc*/){
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::BX(int /*cc*/, int /*Rn*/){
- NOT_IMPLEMENTED(); //Not required
- }
- // ----------------------------------------------------------------------------
- // data transfer...
- // ----------------------------------------------------------------------------
- enum dataTransferOp
- {
- opLDR,opLDRB,opLDRH,opSTR,opSTRB,opSTRH
- };
- void ArmToArm64Assembler::dataTransfer(int op, int cc,
- int Rd, int Rn, uint32_t op_type, uint32_t size)
- {
- const int XSP = 31;
- if(Rn == SP)
- Rn = XSP;
- if(op_type == OPERAND_IMM)
- {
- int addrReg;
- int imm = mAddrMode.immediate;
- if(imm >= 0 && imm < (1<<12))
- *mPC++ = A64_ADD_IMM_X(mTmpReg1, mZeroReg, imm, 0);
- else if(imm < 0 && -imm < (1<<12))
- *mPC++ = A64_SUB_IMM_X(mTmpReg1, mZeroReg, -imm, 0);
- else
- {
- NOT_IMPLEMENTED();
- return;
- }
- addrReg = Rn;
- if(mAddrMode.preindex == true || mAddrMode.postindex == true)
- {
- *mPC++ = A64_ADD_X(mTmpReg2, addrReg, mTmpReg1);
- if(mAddrMode.preindex == true)
- addrReg = mTmpReg2;
- }
- if(cc != AL)
- *mPC++ = A64_B_COND(cc^1, 8);
- *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, addrReg, mZeroReg);
- if(mAddrMode.writeback == true)
- *mPC++ = A64_CSEL_X(Rn, mTmpReg2, Rn, cc);
- }
- else if(op_type == OPERAND_REG_OFFSET)
- {
- if(cc != AL)
- *mPC++ = A64_B_COND(cc^1, 8);
- *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, Rn, mAddrMode.reg_offset);
- }
- else if(op_type > OPERAND_UNSUPPORTED)
- {
- if(cc != AL)
- *mPC++ = A64_B_COND(cc^1, 8);
- *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, Rn, mZeroReg);
- }
- else
- {
- NOT_IMPLEMENTED(); // Not required
- }
- return;
- }
- void ArmToArm64Assembler::ADDR_LDR(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opLDR, cc, Rd, Rn, op_type, 64);
- }
- void ArmToArm64Assembler::ADDR_STR(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opSTR, cc, Rd, Rn, op_type, 64);
- }
- void ArmToArm64Assembler::LDR(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opLDR, cc, Rd, Rn, op_type);
- }
- void ArmToArm64Assembler::LDRB(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opLDRB, cc, Rd, Rn, op_type);
- }
- void ArmToArm64Assembler::STR(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opSTR, cc, Rd, Rn, op_type);
- }
- void ArmToArm64Assembler::STRB(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opSTRB, cc, Rd, Rn, op_type);
- }
- void ArmToArm64Assembler::LDRH(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opLDRH, cc, Rd, Rn, op_type);
- }
- void ArmToArm64Assembler::LDRSB(int /*cc*/, int /*Rd*/, int /*Rn*/, uint32_t /*offset*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::LDRSH(int /*cc*/, int /*Rd*/, int /*Rn*/, uint32_t /*offset*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::STRH(int cc, int Rd, int Rn, uint32_t op_type)
- {
- return dataTransfer(opSTRH, cc, Rd, Rn, op_type);
- }
- // ----------------------------------------------------------------------------
- // block data transfer...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::LDM(int cc, int dir,
- int Rn, int W, uint32_t reg_list)
- {
- const int XSP = 31;
- if(cc != AL || dir != IA || W == 0 || Rn != SP)
- {
- NOT_IMPLEMENTED();
- return;
- }
- for(int i = 0; i < 32; ++i)
- {
- if((reg_list & (1 << i)))
- {
- int reg = i;
- int size = 16;
- *mPC++ = A64_LDR_IMM_PostIndex(reg, XSP, size);
- }
- }
- }
- void ArmToArm64Assembler::STM(int cc, int dir,
- int Rn, int W, uint32_t reg_list)
- {
- const int XSP = 31;
- if(cc != AL || dir != DB || W == 0 || Rn != SP)
- {
- NOT_IMPLEMENTED();
- return;
- }
- for(int i = 31; i >= 0; --i)
- {
- if((reg_list & (1 << i)))
- {
- int size = -16;
- int reg = i;
- *mPC++ = A64_STR_IMM_PreIndex(reg, XSP, size);
- }
- }
- }
- // ----------------------------------------------------------------------------
- // special...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::SWP(int /*cc*/, int /*Rn*/, int /*Rd*/, int /*Rm*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::SWPB(int /*cc*/, int /*Rn*/, int /*Rd*/, int /*Rm*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::SWI(int /*cc*/, uint32_t /*comment*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- // ----------------------------------------------------------------------------
- // DSP instructions...
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::PLD(int /*Rn*/, uint32_t /*offset*/) {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::CLZ(int /*cc*/, int /*Rd*/, int /*Rm*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::QADD(int /*cc*/, int /*Rd*/, int /*Rm*/, int /*Rn*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::QDADD(int /*cc*/, int /*Rd*/, int /*Rm*/, int /*Rn*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::QSUB(int /*cc*/, int /*Rd*/, int /*Rm*/, int /*Rn*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- void ArmToArm64Assembler::QDSUB(int /*cc*/, int /*Rd*/, int /*Rm*/, int /*Rn*/)
- {
- NOT_IMPLEMENTED(); //Not required
- }
- // ----------------------------------------------------------------------------
- // 16 x 16 multiplication
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::SMUL(int cc, int xy,
- int Rd, int Rm, int Rs)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if (xy & xyTB)
- *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 16, 31);
- else
- *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 0, 15);
- if (xy & xyBT)
- *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 16, 31);
- else
- *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 0, 15);
- *mPC++ = A64_MADD_W(Rd,mTmpReg1,mTmpReg2, mZeroReg);
- }
- // ----------------------------------------------------------------------------
- // 32 x 16 multiplication
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::SMULW(int cc, int y, int Rd, int Rm, int Rs)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if (y & yT)
- *mPC++ = A64_SBFM_W(mTmpReg1, Rs, 16, 31);
- else
- *mPC++ = A64_SBFM_W(mTmpReg1, Rs, 0, 15);
- *mPC++ = A64_SBFM_W(mTmpReg2, Rm, 0, 31);
- *mPC++ = A64_SMADDL(mTmpReg3,mTmpReg1,mTmpReg2, mZeroReg);
- *mPC++ = A64_UBFM_X(Rd,mTmpReg3, 16, 47);
- }
- // ----------------------------------------------------------------------------
- // 16 x 16 multiplication and accumulate
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::SMLA(int cc, int xy, int Rd, int Rm, int Rs, int Rn)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- if(xy != xyBB) { NOT_IMPLEMENTED(); return;} //Not required
- *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 0, 15);
- *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 0, 15);
- *mPC++ = A64_MADD_W(Rd, mTmpReg1, mTmpReg2, Rn);
- }
- void ArmToArm64Assembler::SMLAL(int /*cc*/, int /*xy*/,
- int /*RdHi*/, int /*RdLo*/, int /*Rs*/, int /*Rm*/)
- {
- NOT_IMPLEMENTED(); //Not required
- return;
- }
- void ArmToArm64Assembler::SMLAW(int /*cc*/, int /*y*/,
- int /*Rd*/, int /*Rm*/, int /*Rs*/, int /*Rn*/)
- {
- NOT_IMPLEMENTED(); //Not required
- return;
- }
- // ----------------------------------------------------------------------------
- // Byte/half word extract and extend
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::UXTB16(int cc, int Rd, int Rm, int rotate)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- *mPC++ = A64_EXTR_W(mTmpReg1, Rm, Rm, rotate * 8);
- uint32_t imm = 0x00FF00FF;
- *mPC++ = A64_MOVZ_W(mTmpReg2, imm & 0xFFFF, 0);
- *mPC++ = A64_MOVK_W(mTmpReg2, (imm >> 16) & 0x0000FFFF, 16);
- *mPC++ = A64_AND_W(Rd,mTmpReg1, mTmpReg2);
- }
- // ----------------------------------------------------------------------------
- // Bit manipulation
- // ----------------------------------------------------------------------------
- void ArmToArm64Assembler::UBFX(int cc, int Rd, int Rn, int lsb, int width)
- {
- if(cc != AL){ NOT_IMPLEMENTED(); return;} //Not required
- *mPC++ = A64_UBFM_W(Rd, Rn, lsb, lsb + width - 1);
- }
- // ----------------------------------------------------------------------------
- // Shifters...
- // ----------------------------------------------------------------------------
- int ArmToArm64Assembler::buildImmediate(
- uint32_t immediate, uint32_t& rot, uint32_t& imm)
- {
- rot = 0;
- imm = immediate;
- return 0; // Always true
- }
- bool ArmToArm64Assembler::isValidImmediate(uint32_t immediate)
- {
- uint32_t rot, imm;
- return buildImmediate(immediate, rot, imm) == 0;
- }
- uint32_t ArmToArm64Assembler::imm(uint32_t immediate)
- {
- mAddrMode.immediate = immediate;
- mAddrMode.writeback = false;
- mAddrMode.preindex = false;
- mAddrMode.postindex = false;
- return OPERAND_IMM;
- }
- uint32_t ArmToArm64Assembler::reg_imm(int Rm, int type, uint32_t shift)
- {
- mAddrMode.reg_imm_Rm = Rm;
- mAddrMode.reg_imm_type = type;
- mAddrMode.reg_imm_shift = shift;
- return OPERAND_REG_IMM;
- }
- uint32_t ArmToArm64Assembler::reg_rrx(int /*Rm*/)
- {
- NOT_IMPLEMENTED();
- return OPERAND_UNSUPPORTED;
- }
- uint32_t ArmToArm64Assembler::reg_reg(int /*Rm*/, int /*type*/, int /*Rs*/)
- {
- NOT_IMPLEMENTED(); //Not required
- return OPERAND_UNSUPPORTED;
- }
- // ----------------------------------------------------------------------------
- // Addressing modes...
- // ----------------------------------------------------------------------------
- uint32_t ArmToArm64Assembler::immed12_pre(int32_t immed12, int W)
- {
- mAddrMode.immediate = immed12;
- mAddrMode.writeback = W;
- mAddrMode.preindex = true;
- mAddrMode.postindex = false;
- return OPERAND_IMM;
- }
- uint32_t ArmToArm64Assembler::immed12_post(int32_t immed12)
- {
- mAddrMode.immediate = immed12;
- mAddrMode.writeback = true;
- mAddrMode.preindex = false;
- mAddrMode.postindex = true;
- return OPERAND_IMM;
- }
- uint32_t ArmToArm64Assembler::reg_scale_pre(int Rm, int type,
- uint32_t shift, int W)
- {
- if(type != 0 || shift != 0 || W != 0)
- {
- NOT_IMPLEMENTED(); //Not required
- return OPERAND_UNSUPPORTED;
- }
- else
- {
- mAddrMode.reg_offset = Rm;
- return OPERAND_REG_OFFSET;
- }
- }
- uint32_t ArmToArm64Assembler::reg_scale_post(int /*Rm*/, int /*type*/, uint32_t /*shift*/)
- {
- NOT_IMPLEMENTED(); //Not required
- return OPERAND_UNSUPPORTED;
- }
- uint32_t ArmToArm64Assembler::immed8_pre(int32_t immed8, int W)
- {
- mAddrMode.immediate = immed8;
- mAddrMode.writeback = W;
- mAddrMode.preindex = true;
- mAddrMode.postindex = false;
- return OPERAND_IMM;
- }
- uint32_t ArmToArm64Assembler::immed8_post(int32_t immed8)
- {
- mAddrMode.immediate = immed8;
- mAddrMode.writeback = true;
- mAddrMode.preindex = false;
- mAddrMode.postindex = true;
- return OPERAND_IMM;
- }
- uint32_t ArmToArm64Assembler::reg_pre(int Rm, int W)
- {
- if(W != 0)
- {
- NOT_IMPLEMENTED(); //Not required
- return OPERAND_UNSUPPORTED;
- }
- else
- {
- mAddrMode.reg_offset = Rm;
- return OPERAND_REG_OFFSET;
- }
- }
- uint32_t ArmToArm64Assembler::reg_post(int /*Rm*/)
- {
- NOT_IMPLEMENTED(); //Not required
- return OPERAND_UNSUPPORTED;
- }
- // ----------------------------------------------------------------------------
- // A64 instructions
- // ----------------------------------------------------------------------------
- static __unused const char * dataTransferOpName[] =
- {
- "LDR","LDRB","LDRH","STR","STRB","STRH"
- };
- static const uint32_t dataTransferOpCode [] =
- {
- ((0xB8u << 24) | (0x3 << 21) | (0x6 << 13) | (0x0 << 12) |(0x1 << 11)),
- ((0x38u << 24) | (0x3 << 21) | (0x6 << 13) | (0x1 << 12) |(0x1 << 11)),
- ((0x78u << 24) | (0x3 << 21) | (0x6 << 13) | (0x0 << 12) |(0x1 << 11)),
- ((0xB8u << 24) | (0x1 << 21) | (0x6 << 13) | (0x0 << 12) |(0x1 << 11)),
- ((0x38u << 24) | (0x1 << 21) | (0x6 << 13) | (0x1 << 12) |(0x1 << 11)),
- ((0x78u << 24) | (0x1 << 21) | (0x6 << 13) | (0x0 << 12) |(0x1 << 11))
- };
- uint32_t ArmToArm64Assembler::A64_LDRSTR_Wm_SXTW_0(uint32_t op,
- uint32_t size, uint32_t Rt,
- uint32_t Rn, uint32_t Rm)
- {
- if(size == 32)
- {
- LOG_INSTR("%s W%d, [X%d, W%d, SXTW #0]\n",
- dataTransferOpName[op], Rt, Rn, Rm);
- return(dataTransferOpCode[op] | (Rm << 16) | (Rn << 5) | Rt);
- }
- else
- {
- LOG_INSTR("%s X%d, [X%d, W%d, SXTW #0]\n",
- dataTransferOpName[op], Rt, Rn, Rm);
- return(dataTransferOpCode[op] | (0x1<<30) | (Rm<<16) | (Rn<<5)|Rt);
- }
- }
- uint32_t ArmToArm64Assembler::A64_STR_IMM_PreIndex(uint32_t Rt,
- uint32_t Rn, int32_t simm)
- {
- if(Rn == 31)
- LOG_INSTR("STR W%d, [SP, #%d]!\n", Rt, simm);
- else
- LOG_INSTR("STR W%d, [X%d, #%d]!\n", Rt, Rn, simm);
- uint32_t imm9 = (unsigned)(simm) & 0x01FF;
- return (0xB8 << 24) | (imm9 << 12) | (0x3 << 10) | (Rn << 5) | Rt;
- }
- uint32_t ArmToArm64Assembler::A64_LDR_IMM_PostIndex(uint32_t Rt,
- uint32_t Rn, int32_t simm)
- {
- if(Rn == 31)
- LOG_INSTR("LDR W%d, [SP], #%d\n",Rt,simm);
- else
- LOG_INSTR("LDR W%d, [X%d], #%d\n",Rt, Rn, simm);
- uint32_t imm9 = (unsigned)(simm) & 0x01FF;
- return (0xB8 << 24) | (0x1 << 22) |
- (imm9 << 12) | (0x1 << 10) | (Rn << 5) | Rt;
- }
- uint32_t ArmToArm64Assembler::A64_ADD_X_Wm_SXTW(uint32_t Rd,
- uint32_t Rn,
- uint32_t Rm,
- uint32_t amount)
- {
- LOG_INSTR("ADD X%d, X%d, W%d, SXTW #%d\n", Rd, Rn, Rm, amount);
- return ((0x8B << 24) | (0x1 << 21) |(Rm << 16) |
- (0x6 << 13) | (amount << 10) | (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_SUB_X_Wm_SXTW(uint32_t Rd,
- uint32_t Rn,
- uint32_t Rm,
- uint32_t amount)
- {
- LOG_INSTR("SUB X%d, X%d, W%d, SXTW #%d\n", Rd, Rn, Rm, amount);
- return ((0xCB << 24) | (0x1 << 21) |(Rm << 16) |
- (0x6 << 13) | (amount << 10) | (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_B_COND(uint32_t cc, uint32_t offset)
- {
- LOG_INSTR("B.%s #.+%d\n", cc_codes[cc], offset);
- return (0x54 << 24) | ((offset/4) << 5) | (cc);
- }
- uint32_t ArmToArm64Assembler::A64_ADD_X(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount)
- {
- LOG_INSTR("ADD X%d, X%d, X%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x8B << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_ADD_IMM_X(uint32_t Rd, uint32_t Rn,
- uint32_t imm, uint32_t shift)
- {
- LOG_INSTR("ADD X%d, X%d, #%d, LSL #%d\n", Rd, Rn, imm, shift);
- return (0x91 << 24) | ((shift/12) << 22) | (imm << 10) | (Rn << 5) | Rd;
- }
- uint32_t ArmToArm64Assembler::A64_SUB_IMM_X(uint32_t Rd, uint32_t Rn,
- uint32_t imm, uint32_t shift)
- {
- LOG_INSTR("SUB X%d, X%d, #%d, LSL #%d\n", Rd, Rn, imm, shift);
- return (0xD1 << 24) | ((shift/12) << 22) | (imm << 10) | (Rn << 5) | Rd;
- }
- uint32_t ArmToArm64Assembler::A64_ADD_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount)
- {
- LOG_INSTR("ADD W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x0B << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_SUB_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount,
- uint32_t setflag)
- {
- if(setflag == 0)
- {
- LOG_INSTR("SUB W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x4B << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- else
- {
- LOG_INSTR("SUBS W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x6B << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- }
- uint32_t ArmToArm64Assembler::A64_AND_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount)
- {
- LOG_INSTR("AND W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x0A << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_ORR_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount)
- {
- LOG_INSTR("ORR W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x2A << 24) | (shift << 22) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_ORN_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t shift,
- uint32_t amount)
- {
- LOG_INSTR("ORN W%d, W%d, W%d, %s #%d\n",
- Rd, Rn, Rm, shift_codes[shift], amount);
- return ((0x2A << 24) | (shift << 22) | (0x1 << 21) | ( Rm << 16) |
- (amount << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_CSEL_X(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t cond)
- {
- LOG_INSTR("CSEL X%d, X%d, X%d, %s\n", Rd, Rn, Rm, cc_codes[cond]);
- return ((0x9A << 24)|(0x1 << 23)|(Rm << 16) |(cond << 12)| (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_CSEL_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t cond)
- {
- LOG_INSTR("CSEL W%d, W%d, W%d, %s\n", Rd, Rn, Rm, cc_codes[cond]);
- return ((0x1A << 24)|(0x1 << 23)|(Rm << 16) |(cond << 12)| (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_RET(uint32_t Rn)
- {
- LOG_INSTR("RET X%d\n", Rn);
- return ((0xD6 << 24) | (0x1 << 22) | (0x1F << 16) | (Rn << 5));
- }
- uint32_t ArmToArm64Assembler::A64_MOVZ_X(uint32_t Rd, uint32_t imm,
- uint32_t shift)
- {
- LOG_INSTR("MOVZ X%d, #0x%x, LSL #%d\n", Rd, imm, shift);
- return(0xD2 << 24) | (0x1 << 23) | ((shift/16) << 21) | (imm << 5) | Rd;
- }
- uint32_t ArmToArm64Assembler::A64_MOVK_W(uint32_t Rd, uint32_t imm,
- uint32_t shift)
- {
- LOG_INSTR("MOVK W%d, #0x%x, LSL #%d\n", Rd, imm, shift);
- return (0x72 << 24) | (0x1 << 23) | ((shift/16) << 21) | (imm << 5) | Rd;
- }
- uint32_t ArmToArm64Assembler::A64_MOVZ_W(uint32_t Rd, uint32_t imm,
- uint32_t shift)
- {
- LOG_INSTR("MOVZ W%d, #0x%x, LSL #%d\n", Rd, imm, shift);
- return(0x52 << 24) | (0x1 << 23) | ((shift/16) << 21) | (imm << 5) | Rd;
- }
- uint32_t ArmToArm64Assembler::A64_SMADDL(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t Ra)
- {
- LOG_INSTR("SMADDL X%d, W%d, W%d, X%d\n",Rd, Rn, Rm, Ra);
- return ((0x9B << 24) | (0x1 << 21) | (Rm << 16)|(Ra << 10)|(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_MADD_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t Ra)
- {
- LOG_INSTR("MADD W%d, W%d, W%d, W%d\n",Rd, Rn, Rm, Ra);
- return ((0x1B << 24) | (Rm << 16) | (Ra << 10) |(Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_SBFM_W(uint32_t Rd, uint32_t Rn,
- uint32_t immr, uint32_t imms)
- {
- LOG_INSTR("SBFM W%d, W%d, #%d, #%d\n", Rd, Rn, immr, imms);
- return ((0x13 << 24) | (immr << 16) | (imms << 10) | (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_UBFM_W(uint32_t Rd, uint32_t Rn,
- uint32_t immr, uint32_t imms)
- {
- LOG_INSTR("UBFM W%d, W%d, #%d, #%d\n", Rd, Rn, immr, imms);
- return ((0x53 << 24) | (immr << 16) | (imms << 10) | (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_UBFM_X(uint32_t Rd, uint32_t Rn,
- uint32_t immr, uint32_t imms)
- {
- LOG_INSTR("UBFM X%d, X%d, #%d, #%d\n", Rd, Rn, immr, imms);
- return ((0xD3 << 24) | (0x1 << 22) |
- (immr << 16) | (imms << 10) | (Rn << 5) | Rd);
- }
- uint32_t ArmToArm64Assembler::A64_EXTR_W(uint32_t Rd, uint32_t Rn,
- uint32_t Rm, uint32_t lsb)
- {
- LOG_INSTR("EXTR W%d, W%d, W%d, #%d\n", Rd, Rn, Rm, lsb);
- return (0x13 << 24)|(0x1 << 23) | (Rm << 16) | (lsb << 10)|(Rn << 5) | Rd;
- }
- }; // namespace android
|