/* TILE-Gx opcode information. * * Copyright 2011 Tilera Corporation. All Rights Reserved. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation, version 2. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or * NON INFRINGEMENT. See the GNU General Public License for * more details. * * * * * */ #ifndef __ARCH_OPCODE_H__ #define __ARCH_OPCODE_H__ #ifndef __ASSEMBLER__ typedef uint64_t tilegx_bundle_bits; /* These are the bits that determine if a bundle is in the X encoding. */ #define TILEGX_BUNDLE_MODE_MASK ((tilegx_bundle_bits)3 << 62) enum { /* Maximum number of instructions in a bundle (2 for X, 3 for Y). */ TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE = 3, /* How many different pipeline encodings are there? X0, X1, Y0, Y1, Y2. */ TILEGX_NUM_PIPELINE_ENCODINGS = 5, /* Log base 2 of TILEGX_BUNDLE_SIZE_IN_BYTES. */ TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES = 3, /* Instructions take this many bytes. */ TILEGX_BUNDLE_SIZE_IN_BYTES = 1 << TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES, /* Log base 2 of TILEGX_BUNDLE_ALIGNMENT_IN_BYTES. */ TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES = 3, /* Bundles should be aligned modulo this number of bytes. */ TILEGX_BUNDLE_ALIGNMENT_IN_BYTES = (1 << TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES), /* Number of registers (some are magic, such as network I/O). */ TILEGX_NUM_REGISTERS = 64, }; /* Make a few "tile_" variables to simplify common code between architectures. */ typedef tilegx_bundle_bits tile_bundle_bits; #define TILE_BUNDLE_SIZE_IN_BYTES TILEGX_BUNDLE_SIZE_IN_BYTES #define TILE_BUNDLE_ALIGNMENT_IN_BYTES TILEGX_BUNDLE_ALIGNMENT_IN_BYTES #define TILE_LOG2_BUNDLE_ALIGNMENT_IN_BYTES \ TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES #define TILE_BPT_BUNDLE TILEGX_BPT_BUNDLE /* 64-bit pattern for a { bpt ; nop } bundle. */ #define TILEGX_BPT_BUNDLE 0x286a44ae51485000ULL static inline unsigned int get_BFEnd_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_BFOpcodeExtension_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 24)) & 0xf); } static inline unsigned int get_BFStart_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 18)) & 0x3f); } static inline unsigned int get_BrOff_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x0000003f) | (((unsigned int)(n >> 37)) & 0x0001ffc0); } static inline unsigned int get_BrType_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 54)) & 0x1f); } static inline unsigned int get_Dest_Imm8_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x0000003f) | (((unsigned int)(n >> 43)) & 0x000000c0); } static inline unsigned int get_Dest_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 0)) & 0x3f); } static inline unsigned int get_Dest_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x3f); } static inline unsigned int get_Dest_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 0)) & 0x3f); } static inline unsigned int get_Dest_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x3f); } static inline unsigned int get_Imm16_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0xffff); } static inline unsigned int get_Imm16_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0xffff); } static inline unsigned int get_Imm8OpcodeExtension_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 20)) & 0xff); } static inline unsigned int get_Imm8OpcodeExtension_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 51)) & 0xff); } static inline unsigned int get_Imm8_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0xff); } static inline unsigned int get_Imm8_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0xff); } static inline unsigned int get_Imm8_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0xff); } static inline unsigned int get_Imm8_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0xff); } static inline unsigned int get_JumpOff_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x7ffffff); } static inline unsigned int get_JumpOpcodeExtension_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 58)) & 0x1); } static inline unsigned int get_MF_Imm14_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 37)) & 0x3fff); } static inline unsigned int get_MT_Imm14_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 31)) & 0x0000003f) | (((unsigned int)(n >> 37)) & 0x00003fc0); } static inline unsigned int get_Mode(tilegx_bundle_bits n) { return (((unsigned int)(n >> 62)) & 0x3); } static inline unsigned int get_Opcode_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 28)) & 0x7); } static inline unsigned int get_Opcode_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 59)) & 0x7); } static inline unsigned int get_Opcode_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 27)) & 0xf); } static inline unsigned int get_Opcode_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 58)) & 0xf); } static inline unsigned int get_Opcode_Y2(tilegx_bundle_bits n) { return (((n >> 26)) & 0x00000001) | (((unsigned int)(n >> 56)) & 0x00000002); } static inline unsigned int get_RRROpcodeExtension_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 18)) & 0x3ff); } static inline unsigned int get_RRROpcodeExtension_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 49)) & 0x3ff); } static inline unsigned int get_RRROpcodeExtension_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 18)) & 0x3); } static inline unsigned int get_RRROpcodeExtension_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 49)) & 0x3); } static inline unsigned int get_ShAmt_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_ShAmt_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline unsigned int get_ShAmt_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_ShAmt_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline unsigned int get_ShiftOpcodeExtension_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 18)) & 0x3ff); } static inline unsigned int get_ShiftOpcodeExtension_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 49)) & 0x3ff); } static inline unsigned int get_ShiftOpcodeExtension_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 18)) & 0x3); } static inline unsigned int get_ShiftOpcodeExtension_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 49)) & 0x3); } static inline unsigned int get_SrcA_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 6)) & 0x3f); } static inline unsigned int get_SrcA_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 37)) & 0x3f); } static inline unsigned int get_SrcA_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 6)) & 0x3f); } static inline unsigned int get_SrcA_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 37)) & 0x3f); } static inline unsigned int get_SrcA_Y2(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 20)) & 0x3f); } static inline unsigned int get_SrcBDest_Y2(tilegx_bundle_bits n) { return (((unsigned int)(n >> 51)) & 0x3f); } static inline unsigned int get_SrcB_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_SrcB_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline unsigned int get_SrcB_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_SrcB_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline unsigned int get_UnaryOpcodeExtension_X0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_UnaryOpcodeExtension_X1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline unsigned int get_UnaryOpcodeExtension_Y0(tilegx_bundle_bits num) { const unsigned int n = (unsigned int)num; return (((n >> 12)) & 0x3f); } static inline unsigned int get_UnaryOpcodeExtension_Y1(tilegx_bundle_bits n) { return (((unsigned int)(n >> 43)) & 0x3f); } static inline int sign_extend(int n, int num_bits) { int shift = (int)(sizeof(int) * 8 - num_bits); return (n << shift) >> shift; } static inline tilegx_bundle_bits create_BFEnd_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_BFOpcodeExtension_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xf) << 24); } static inline tilegx_bundle_bits create_BFStart_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 18); } static inline tilegx_bundle_bits create_BrOff_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) | (((tilegx_bundle_bits)(n & 0x0001ffc0)) << 37); } static inline tilegx_bundle_bits create_BrType_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x1f)) << 54); } static inline tilegx_bundle_bits create_Dest_Imm8_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) | (((tilegx_bundle_bits)(n & 0x000000c0)) << 43); } static inline tilegx_bundle_bits create_Dest_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 0); } static inline tilegx_bundle_bits create_Dest_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 31); } static inline tilegx_bundle_bits create_Dest_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 0); } static inline tilegx_bundle_bits create_Dest_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 31); } static inline tilegx_bundle_bits create_Imm16_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xffff) << 12); } static inline tilegx_bundle_bits create_Imm16_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0xffff)) << 43); } static inline tilegx_bundle_bits create_Imm8OpcodeExtension_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xff) << 20); } static inline tilegx_bundle_bits create_Imm8OpcodeExtension_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0xff)) << 51); } static inline tilegx_bundle_bits create_Imm8_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xff) << 12); } static inline tilegx_bundle_bits create_Imm8_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0xff)) << 43); } static inline tilegx_bundle_bits create_Imm8_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xff) << 12); } static inline tilegx_bundle_bits create_Imm8_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0xff)) << 43); } static inline tilegx_bundle_bits create_JumpOff_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x7ffffff)) << 31); } static inline tilegx_bundle_bits create_JumpOpcodeExtension_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x1)) << 58); } static inline tilegx_bundle_bits create_MF_Imm14_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3fff)) << 37); } static inline tilegx_bundle_bits create_MT_Imm14_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) | (((tilegx_bundle_bits)(n & 0x00003fc0)) << 37); } static inline tilegx_bundle_bits create_Mode(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3)) << 62); } static inline tilegx_bundle_bits create_Opcode_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x7) << 28); } static inline tilegx_bundle_bits create_Opcode_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x7)) << 59); } static inline tilegx_bundle_bits create_Opcode_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0xf) << 27); } static inline tilegx_bundle_bits create_Opcode_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0xf)) << 58); } static inline tilegx_bundle_bits create_Opcode_Y2(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x00000001) << 26) | (((tilegx_bundle_bits)(n & 0x00000002)) << 56); } static inline tilegx_bundle_bits create_RRROpcodeExtension_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3ff) << 18); } static inline tilegx_bundle_bits create_RRROpcodeExtension_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3ff)) << 49); } static inline tilegx_bundle_bits create_RRROpcodeExtension_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3) << 18); } static inline tilegx_bundle_bits create_RRROpcodeExtension_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3)) << 49); } static inline tilegx_bundle_bits create_ShAmt_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_ShAmt_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } static inline tilegx_bundle_bits create_ShAmt_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_ShAmt_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } static inline tilegx_bundle_bits create_ShiftOpcodeExtension_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3ff) << 18); } static inline tilegx_bundle_bits create_ShiftOpcodeExtension_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3ff)) << 49); } static inline tilegx_bundle_bits create_ShiftOpcodeExtension_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3) << 18); } static inline tilegx_bundle_bits create_ShiftOpcodeExtension_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3)) << 49); } static inline tilegx_bundle_bits create_SrcA_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 6); } static inline tilegx_bundle_bits create_SrcA_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 37); } static inline tilegx_bundle_bits create_SrcA_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 6); } static inline tilegx_bundle_bits create_SrcA_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 37); } static inline tilegx_bundle_bits create_SrcA_Y2(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 20); } static inline tilegx_bundle_bits create_SrcBDest_Y2(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 51); } static inline tilegx_bundle_bits create_SrcB_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_SrcB_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } static inline tilegx_bundle_bits create_SrcB_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_SrcB_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } static inline tilegx_bundle_bits create_UnaryOpcodeExtension_X0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_UnaryOpcodeExtension_X1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } static inline tilegx_bundle_bits create_UnaryOpcodeExtension_Y0(int num) { const unsigned int n = (unsigned int)num; return ((n & 0x3f) << 12); } static inline tilegx_bundle_bits create_UnaryOpcodeExtension_Y1(int num) { const unsigned int n = (unsigned int)num; return (((tilegx_bundle_bits)(n & 0x3f)) << 43); } enum { ADDI_IMM8_OPCODE_X0 = 1, ADDI_IMM8_OPCODE_X1 = 1, ADDI_OPCODE_Y0 = 0, ADDI_OPCODE_Y1 = 1, ADDLI_OPCODE_X0 = 1, ADDLI_OPCODE_X1 = 0, ADDXI_IMM8_OPCODE_X0 = 2, ADDXI_IMM8_OPCODE_X1 = 2, ADDXI_OPCODE_Y0 = 1, ADDXI_OPCODE_Y1 = 2, ADDXLI_OPCODE_X0 = 2, ADDXLI_OPCODE_X1 = 1, ADDXSC_RRR_0_OPCODE_X0 = 1, ADDXSC_RRR_0_OPCODE_X1 = 1, ADDX_RRR_0_OPCODE_X0 = 2, ADDX_RRR_0_OPCODE_X1 = 2, ADDX_RRR_0_OPCODE_Y0 = 0, ADDX_RRR_0_OPCODE_Y1 = 0, ADD_RRR_0_OPCODE_X0 = 3, ADD_RRR_0_OPCODE_X1 = 3, ADD_RRR_0_OPCODE_Y0 = 1, ADD_RRR_0_OPCODE_Y1 = 1, ANDI_IMM8_OPCODE_X0 = 3, ANDI_IMM8_OPCODE_X1 = 3, ANDI_OPCODE_Y0 = 2, ANDI_OPCODE_Y1 = 3, AND_RRR_0_OPCODE_X0 = 4, AND_RRR_0_OPCODE_X1 = 4, AND_RRR_5_OPCODE_Y0 = 0, AND_RRR_5_OPCODE_Y1 = 0, BEQZT_BRANCH_OPCODE_X1 = 16, BEQZ_BRANCH_OPCODE_X1 = 17, BFEXTS_BF_OPCODE_X0 = 4, BFEXTU_BF_OPCODE_X0 = 5, BFINS_BF_OPCODE_X0 = 6, BF_OPCODE_X0 = 3, BGEZT_BRANCH_OPCODE_X1 = 18, BGEZ_BRANCH_OPCODE_X1 = 19, BGTZT_BRANCH_OPCODE_X1 = 20, BGTZ_BRANCH_OPCODE_X1 = 21, BLBCT_BRANCH_OPCODE_X1 = 22, BLBC_BRANCH_OPCODE_X1 = 23, BLBST_BRANCH_OPCODE_X1 = 24, BLBS_BRANCH_OPCODE_X1 = 25, BLEZT_BRANCH_OPCODE_X1 = 26, BLEZ_BRANCH_OPCODE_X1 = 27, BLTZT_BRANCH_OPCODE_X1 = 28, BLTZ_BRANCH_OPCODE_X1 = 29, BNEZT_BRANCH_OPCODE_X1 = 30, BNEZ_BRANCH_OPCODE_X1 = 31, BRANCH_OPCODE_X1 = 2, CMOVEQZ_RRR_0_OPCODE_X0 = 5, CMOVEQZ_RRR_4_OPCODE_Y0 = 0, CMOVNEZ_RRR_0_OPCODE_X0 = 6, CMOVNEZ_RRR_4_OPCODE_Y0 = 1, CMPEQI_IMM8_OPCODE_X0 = 4, CMPEQI_IMM8_OPCODE_X1 = 4, CMPEQI_OPCODE_Y0 = 3, CMPEQI_OPCODE_Y1 = 4, CMPEQ_RRR_0_OPCODE_X0 = 7, CMPEQ_RRR_0_OPCODE_X1 = 5, CMPEQ_RRR_3_OPCODE_Y0 = 0, CMPEQ_RRR_3_OPCODE_Y1 = 2, CMPEXCH4_RRR_0_OPCODE_X1 = 6, CMPEXCH_RRR_0_OPCODE_X1 = 7, CMPLES_RRR_0_OPCODE_X0 = 8, CMPLES_RRR_0_OPCODE_X1 = 8, CMPLES_RRR_2_OPCODE_Y0 = 0, CMPLES_RRR_2_OPCODE_Y1 = 0, CMPLEU_RRR_0_OPCODE_X0 = 9, CMPLEU_RRR_0_OPCODE_X1 = 9, CMPLEU_RRR_2_OPCODE_Y0 = 1, CMPLEU_RRR_2_OPCODE_Y1 = 1, CMPLTSI_IMM8_OPCODE_X0 = 5, CMPLTSI_IMM8_OPCODE_X1 = 5, CMPLTSI_OPCODE_Y0 = 4, CMPLTSI_OPCODE_Y1 = 5, CMPLTS_RRR_0_OPCODE_X0 = 10, CMPLTS_RRR_0_OPCODE_X1 = 10, CMPLTS_RRR_2_OPCODE_Y0 = 2, CMPLTS_RRR_2_OPCODE_Y1 = 2, CMPLTUI_IMM8_OPCODE_X0 = 6, CMPLTUI_IMM8_OPCODE_X1 = 6, CMPLTU_RRR_0_OPCODE_X0 = 11, CMPLTU_RRR_0_OPCODE_X1 = 11, CMPLTU_RRR_2_OPCODE_Y0 = 3, CMPLTU_RRR_2_OPCODE_Y1 = 3, CMPNE_RRR_0_OPCODE_X0 = 12, CMPNE_RRR_0_OPCODE_X1 = 12, CMPNE_RRR_3_OPCODE_Y0 = 1, CMPNE_RRR_3_OPCODE_Y1 = 3, CMULAF_RRR_0_OPCODE_X0 = 13, CMULA_RRR_0_OPCODE_X0 = 14, CMULFR_RRR_0_OPCODE_X0 = 15, CMULF_RRR_0_OPCODE_X0 = 16, CMULHR_RRR_0_OPCODE_X0 = 17, CMULH_RRR_0_OPCODE_X0 = 18, CMUL_RRR_0_OPCODE_X0 = 19, CNTLZ_UNARY_OPCODE_X0 = 1, CNTLZ_UNARY_OPCODE_Y0 = 1, CNTTZ_UNARY_OPCODE_X0 = 2, CNTTZ_UNARY_OPCODE_Y0 = 2, CRC32_32_RRR_0_OPCODE_X0 = 20, CRC32_8_RRR_0_OPCODE_X0 = 21, DBLALIGN2_RRR_0_OPCODE_X0 = 22, DBLALIGN2_RRR_0_OPCODE_X1 = 13, DBLALIGN4_RRR_0_OPCODE_X0 = 23, DBLALIGN4_RRR_0_OPCODE_X1 = 14, DBLALIGN6_RRR_0_OPCODE_X0 = 24, DBLALIGN6_RRR_0_OPCODE_X1 = 15, DBLALIGN_RRR_0_OPCODE_X0 = 25, DRAIN_UNARY_OPCODE_X1 = 1, DTLBPR_UNARY_OPCODE_X1 = 2, EXCH4_RRR_0_OPCODE_X1 = 16, EXCH_RRR_0_OPCODE_X1 = 17, FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 = 26, FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 = 27, FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 = 28, FDOUBLE_PACK1_RRR_0_OPCODE_X0 = 29, FDOUBLE_PACK2_RRR_0_OPCODE_X0 = 30, FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 = 31, FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 = 32, FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 = 33, FETCHADD4_RRR_0_OPCODE_X1 = 18, FETCHADDGEZ4_RRR_0_OPCODE_X1 = 19, FETCHADDGEZ_RRR_0_OPCODE_X1 = 20, FETCHADD_RRR_0_OPCODE_X1 = 21, FETCHAND4_RRR_0_OPCODE_X1 = 22, FETCHAND_RRR_0_OPCODE_X1 = 23, FETCHOR4_RRR_0_OPCODE_X1 = 24, FETCHOR_RRR_0_OPCODE_X1 = 25, FINV_UNARY_OPCODE_X1 = 3, FLUSHWB_UNARY_OPCODE_X1 = 4, FLUSH_UNARY_OPCODE_X1 = 5, FNOP_UNARY_OPCODE_X0 = 3, FNOP_UNARY_OPCODE_X1 = 6, FNOP_UNARY_OPCODE_Y0 = 3, FNOP_UNARY_OPCODE_Y1 = 8, FSINGLE_ADD1_RRR_0_OPCODE_X0 = 34, FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 = 35, FSINGLE_MUL1_RRR_0_OPCODE_X0 = 36, FSINGLE_MUL2_RRR_0_OPCODE_X0 = 37, FSINGLE_PACK1_UNARY_OPCODE_X0 = 4, FSINGLE_PACK1_UNARY_OPCODE_Y0 = 4, FSINGLE_PACK2_RRR_0_OPCODE_X0 = 38, FSINGLE_SUB1_RRR_0_OPCODE_X0 = 39, ICOH_UNARY_OPCODE_X1 = 7, ILL_UNARY_OPCODE_X1 = 8, ILL_UNARY_OPCODE_Y1 = 9, IMM8_OPCODE_X0 = 4, IMM8_OPCODE_X1 = 3, INV_UNARY_OPCODE_X1 = 9, IRET_UNARY_OPCODE_X1 = 10, JALRP_UNARY_OPCODE_X1 = 11, JALRP_UNARY_OPCODE_Y1 = 10, JALR_UNARY_OPCODE_X1 = 12, JALR_UNARY_OPCODE_Y1 = 11, JAL_JUMP_OPCODE_X1 = 0, JRP_UNARY_OPCODE_X1 = 13, JRP_UNARY_OPCODE_Y1 = 12, JR_UNARY_OPCODE_X1 = 14, JR_UNARY_OPCODE_Y1 = 13, JUMP_OPCODE_X1 = 4, J_JUMP_OPCODE_X1 = 1, LD1S_ADD_IMM8_OPCODE_X1 = 7, LD1S_OPCODE_Y2 = 0, LD1S_UNARY_OPCODE_X1 = 15, LD1U_ADD_IMM8_OPCODE_X1 = 8, LD1U_OPCODE_Y2 = 1, LD1U_UNARY_OPCODE_X1 = 16, LD2S_ADD_IMM8_OPCODE_X1 = 9, LD2S_OPCODE_Y2 = 2, LD2S_UNARY_OPCODE_X1 = 17, LD2U_ADD_IMM8_OPCODE_X1 = 10, LD2U_OPCODE_Y2 = 3, LD2U_UNARY_OPCODE_X1 = 18, LD4S_ADD_IMM8_OPCODE_X1 = 11, LD4S_OPCODE_Y2 = 1, LD4S_UNARY_OPCODE_X1 = 19, LD4U_ADD_IMM8_OPCODE_X1 = 12, LD4U_OPCODE_Y2 = 2, LD4U_UNARY_OPCODE_X1 = 20, LDNA_UNARY_OPCODE_X1 = 21, LDNT1S_ADD_IMM8_OPCODE_X1 = 13, LDNT1S_UNARY_OPCODE_X1 = 22, LDNT1U_ADD_IMM8_OPCODE_X1 = 14, LDNT1U_UNARY_OPCODE_X1 = 23, LDNT2S_ADD_IMM8_OPCODE_X1 = 15, LDNT2S_UNARY_OPCODE_X1 = 24, LDNT2U_ADD_IMM8_OPCODE_X1 = 16, LDNT2U_UNARY_OPCODE_X1 = 25, LDNT4S_ADD_IMM8_OPCODE_X1 = 17, LDNT4S_UNARY_OPCODE_X1 = 26, LDNT4U_ADD_IMM8_OPCODE_X1 = 18, LDNT4U_UNARY_OPCODE_X1 = 27, LDNT_ADD_IMM8_OPCODE_X1 = 19, LDNT_UNARY_OPCODE_X1 = 28, LD_ADD_IMM8_OPCODE_X1 = 20, LD_OPCODE_Y2 = 3, LD_UNARY_OPCODE_X1 = 29, LNK_UNARY_OPCODE_X1 = 30, LNK_UNARY_OPCODE_Y1 = 14, LDNA_ADD_IMM8_OPCODE_X1 = 21, MFSPR_IMM8_OPCODE_X1 = 22, MF_UNARY_OPCODE_X1 = 31, MM_BF_OPCODE_X0 = 7, MNZ_RRR_0_OPCODE_X0 = 40, MNZ_RRR_0_OPCODE_X1 = 26, MNZ_RRR_4_OPCODE_Y0 = 2, MNZ_RRR_4_OPCODE_Y1 = 2, MODE_OPCODE_YA2 = 1, MODE_OPCODE_YB2 = 2, MODE_OPCODE_YC2 = 3, MTSPR_IMM8_OPCODE_X1 = 23, MULAX_RRR_0_OPCODE_X0 = 41, MULAX_RRR_3_OPCODE_Y0 = 2, MULA_HS_HS_RRR_0_OPCODE_X0 = 42, MULA_HS_HS_RRR_9_OPCODE_Y0 = 0, MULA_HS_HU_RRR_0_OPCODE_X0 = 43, MULA_HS_LS_RRR_0_OPCODE_X0 = 44, MULA_HS_LU_RRR_0_OPCODE_X0 = 45, MULA_HU_HU_RRR_0_OPCODE_X0 = 46, MULA_HU_HU_RRR_9_OPCODE_Y0 = 1, MULA_HU_LS_RRR_0_OPCODE_X0 = 47, MULA_HU_LU_RRR_0_OPCODE_X0 = 48, MULA_LS_LS_RRR_0_OPCODE_X0 = 49, MULA_LS_LS_RRR_9_OPCODE_Y0 = 2, MULA_LS_LU_RRR_0_OPCODE_X0 = 50, MULA_LU_LU_RRR_0_OPCODE_X0 = 51, MULA_LU_LU_RRR_9_OPCODE_Y0 = 3, MULX_RRR_0_OPCODE_X0 = 52, MULX_RRR_3_OPCODE_Y0 = 3, MUL_HS_HS_RRR_0_OPCODE_X0 = 53, MUL_HS_HS_RRR_8_OPCODE_Y0 = 0, MUL_HS_HU_RRR_0_OPCODE_X0 = 54, MUL_HS_LS_RRR_0_OPCODE_X0 = 55, MUL_HS_LU_RRR_0_OPCODE_X0 = 56, MUL_HU_HU_RRR_0_OPCODE_X0 = 57, MUL_HU_HU_RRR_8_OPCODE_Y0 = 1, MUL_HU_LS_RRR_0_OPCODE_X0 = 58, MUL_HU_LU_RRR_0_OPCODE_X0 = 59, MUL_LS_LS_RRR_0_OPCODE_X0 = 60, MUL_LS_LS_RRR_8_OPCODE_Y0 = 2, MUL_LS_LU_RRR_0_OPCODE_X0 = 61, MUL_LU_LU_RRR_0_OPCODE_X0 = 62, MUL_LU_LU_RRR_8_OPCODE_Y0 = 3, MZ_RRR_0_OPCODE_X0 = 63, MZ_RRR_0_OPCODE_X1 = 27, MZ_RRR_4_OPCODE_Y0 = 3, MZ_RRR_4_OPCODE_Y1 = 3, NAP_UNARY_OPCODE_X1 = 32, NOP_UNARY_OPCODE_X0 = 5, NOP_UNARY_OPCODE_X1 = 33, NOP_UNARY_OPCODE_Y0 = 5, NOP_UNARY_OPCODE_Y1 = 15, NOR_RRR_0_OPCODE_X0 = 64, NOR_RRR_0_OPCODE_X1 = 28, NOR_RRR_5_OPCODE_Y0 = 1, NOR_RRR_5_OPCODE_Y1 = 1, ORI_IMM8_OPCODE_X0 = 7, ORI_IMM8_OPCODE_X1 = 24, OR_RRR_0_OPCODE_X0 = 65, OR_RRR_0_OPCODE_X1 = 29, OR_RRR_5_OPCODE_Y0 = 2, OR_RRR_5_OPCODE_Y1 = 2, PCNT_UNARY_OPCODE_X0 = 6, PCNT_UNARY_OPCODE_Y0 = 6, REVBITS_UNARY_OPCODE_X0 = 7, REVBITS_UNARY_OPCODE_Y0 = 7, REVBYTES_UNARY_OPCODE_X0 = 8, REVBYTES_UNARY_OPCODE_Y0 = 8, ROTLI_SHIFT_OPCODE_X0 = 1, ROTLI_SHIFT_OPCODE_X1 = 1, ROTLI_SHIFT_OPCODE_Y0 = 0, ROTLI_SHIFT_OPCODE_Y1 = 0, ROTL_RRR_0_OPCODE_X0 = 66, ROTL_RRR_0_OPCODE_X1 = 30, ROTL_RRR_6_OPCODE_Y0 = 0, ROTL_RRR_6_OPCODE_Y1 = 0, RRR_0_OPCODE_X0 = 5, RRR_0_OPCODE_X1 = 5, RRR_0_OPCODE_Y0 = 5, RRR_0_OPCODE_Y1 = 6, RRR_1_OPCODE_Y0 = 6, RRR_1_OPCODE_Y1 = 7, RRR_2_OPCODE_Y0 = 7, RRR_2_OPCODE_Y1 = 8, RRR_3_OPCODE_Y0 = 8, RRR_3_OPCODE_Y1 = 9, RRR_4_OPCODE_Y0 = 9, RRR_4_OPCODE_Y1 = 10, RRR_5_OPCODE_Y0 = 10, RRR_5_OPCODE_Y1 = 11, RRR_6_OPCODE_Y0 = 11, RRR_6_OPCODE_Y1 = 12, RRR_7_OPCODE_Y0 = 12, RRR_7_OPCODE_Y1 = 13, RRR_8_OPCODE_Y0 = 13, RRR_9_OPCODE_Y0 = 14, SHIFT_OPCODE_X0 = 6, SHIFT_OPCODE_X1 = 6, SHIFT_OPCODE_Y0 = 15, SHIFT_OPCODE_Y1 = 14, SHL16INSLI_OPCODE_X0 = 7, SHL16INSLI_OPCODE_X1 = 7, SHL1ADDX_RRR_0_OPCODE_X0 = 67, SHL1ADDX_RRR_0_OPCODE_X1 = 31, SHL1ADDX_RRR_7_OPCODE_Y0 = 1, SHL1ADDX_RRR_7_OPCODE_Y1 = 1, SHL1ADD_RRR_0_OPCODE_X0 = 68, SHL1ADD_RRR_0_OPCODE_X1 = 32, SHL1ADD_RRR_1_OPCODE_Y0 = 0, SHL1ADD_RRR_1_OPCODE_Y1 = 0, SHL2ADDX_RRR_0_OPCODE_X0 = 69, SHL2ADDX_RRR_0_OPCODE_X1 = 33, SHL2ADDX_RRR_7_OPCODE_Y0 = 2, SHL2ADDX_RRR_7_OPCODE_Y1 = 2, SHL2ADD_RRR_0_OPCODE_X0 = 70, SHL2ADD_RRR_0_OPCODE_X1 = 34, SHL2ADD_RRR_1_OPCODE_Y0 = 1, SHL2ADD_RRR_1_OPCODE_Y1 = 1, SHL3ADDX_RRR_0_OPCODE_X0 = 71, SHL3ADDX_RRR_0_OPCODE_X1 = 35, SHL3ADDX_RRR_7_OPCODE_Y0 = 3, SHL3ADDX_RRR_7_OPCODE_Y1 = 3, SHL3ADD_RRR_0_OPCODE_X0 = 72, SHL3ADD_RRR_0_OPCODE_X1 = 36, SHL3ADD_RRR_1_OPCODE_Y0 = 2, SHL3ADD_RRR_1_OPCODE_Y1 = 2, SHLI_SHIFT_OPCODE_X0 = 2, SHLI_SHIFT_OPCODE_X1 = 2, SHLI_SHIFT_OPCODE_Y0 = 1, SHLI_SHIFT_OPCODE_Y1 = 1, SHLXI_SHIFT_OPCODE_X0 = 3, SHLXI_SHIFT_OPCODE_X1 = 3, SHLX_RRR_0_OPCODE_X0 = 73, SHLX_RRR_0_OPCODE_X1 = 37, SHL_RRR_0_OPCODE_X0 = 74, SHL_RRR_0_OPCODE_X1 = 38, SHL_RRR_6_OPCODE_Y0 = 1, SHL_RRR_6_OPCODE_Y1 = 1, SHRSI_SHIFT_OPCODE_X0 = 4, SHRSI_SHIFT_OPCODE_X1 = 4, SHRSI_SHIFT_OPCODE_Y0 = 2, SHRSI_SHIFT_OPCODE_Y1 = 2, SHRS_RRR_0_OPCODE_X0 = 75, SHRS_RRR_0_OPCODE_X1 = 39, SHRS_RRR_6_OPCODE_Y0 = 2, SHRS_RRR_6_OPCODE_Y1 = 2, SHRUI_SHIFT_OPCODE_X0 = 5, SHRUI_SHIFT_OPCODE_X1 = 5, SHRUI_SHIFT_OPCODE_Y0 = 3, SHRUI_SHIFT_OPCODE_Y1 = 3, SHRUXI_SHIFT_OPCODE_X0 = 6, SHRUXI_SHIFT_OPCODE_X1 = 6, SHRUX_RRR_0_OPCODE_X0 = 76, SHRUX_RRR_0_OPCODE_X1 = 40, SHRU_RRR_0_OPCODE_X0 = 77, SHRU_RRR_0_OPCODE_X1 = 41, SHRU_RRR_6_OPCODE_Y0 = 3, SHRU_RRR_6_OPCODE_Y1 = 3, SHUFFLEBYTES_RRR_0_OPCODE_X0 = 78, ST1_ADD_IMM8_OPCODE_X1 = 25, ST1_OPCODE_Y2 = 0, ST1_RRR_0_OPCODE_X1 = 42, ST2_ADD_IMM8_OPCODE_X1 = 26, ST2_OPCODE_Y2 = 1, ST2_RRR_0_OPCODE_X1 = 43, ST4_ADD_IMM8_OPCODE_X1 = 27, ST4_OPCODE_Y2 = 2, ST4_RRR_0_OPCODE_X1 = 44, STNT1_ADD_IMM8_OPCODE_X1 = 28, STNT1_RRR_0_OPCODE_X1 = 45, STNT2_ADD_IMM8_OPCODE_X1 = 29, STNT2_RRR_0_OPCODE_X1 = 46, STNT4_ADD_IMM8_OPCODE_X1 = 30, STNT4_RRR_0_OPCODE_X1 = 47, STNT_ADD_IMM8_OPCODE_X1 = 31, STNT_RRR_0_OPCODE_X1 = 48, ST_ADD_IMM8_OPCODE_X1 = 32, ST_OPCODE_Y2 = 3, ST_RRR_0_OPCODE_X1 = 49, SUBXSC_RRR_0_OPCODE_X0 = 79, SUBXSC_RRR_0_OPCODE_X1 = 50, SUBX_RRR_0_OPCODE_X0 = 80, SUBX_RRR_0_OPCODE_X1 = 51, SUBX_RRR_0_OPCODE_Y0 = 2, SUBX_RRR_0_OPCODE_Y1 = 2, SUB_RRR_0_OPCODE_X0 = 81, SUB_RRR_0_OPCODE_X1 = 52, SUB_RRR_0_OPCODE_Y0 = 3, SUB_RRR_0_OPCODE_Y1 = 3, SWINT0_UNARY_OPCODE_X1 = 34, SWINT1_UNARY_OPCODE_X1 = 35, SWINT2_UNARY_OPCODE_X1 = 36, SWINT3_UNARY_OPCODE_X1 = 37, TBLIDXB0_UNARY_OPCODE_X0 = 9, TBLIDXB0_UNARY_OPCODE_Y0 = 9, TBLIDXB1_UNARY_OPCODE_X0 = 10, TBLIDXB1_UNARY_OPCODE_Y0 = 10, TBLIDXB2_UNARY_OPCODE_X0 = 11, TBLIDXB2_UNARY_OPCODE_Y0 = 11, TBLIDXB3_UNARY_OPCODE_X0 = 12, TBLIDXB3_UNARY_OPCODE_Y0 = 12, UNARY_RRR_0_OPCODE_X0 = 82, UNARY_RRR_0_OPCODE_X1 = 53, UNARY_RRR_1_OPCODE_Y0 = 3, UNARY_RRR_1_OPCODE_Y1 = 3, V1ADDI_IMM8_OPCODE_X0 = 8, V1ADDI_IMM8_OPCODE_X1 = 33, V1ADDUC_RRR_0_OPCODE_X0 = 83, V1ADDUC_RRR_0_OPCODE_X1 = 54, V1ADD_RRR_0_OPCODE_X0 = 84, V1ADD_RRR_0_OPCODE_X1 = 55, V1ADIFFU_RRR_0_OPCODE_X0 = 85, V1AVGU_RRR_0_OPCODE_X0 = 86, V1CMPEQI_IMM8_OPCODE_X0 = 9, V1CMPEQI_IMM8_OPCODE_X1 = 34, V1CMPEQ_RRR_0_OPCODE_X0 = 87, V1CMPEQ_RRR_0_OPCODE_X1 = 56, V1CMPLES_RRR_0_OPCODE_X0 = 88, V1CMPLES_RRR_0_OPCODE_X1 = 57, V1CMPLEU_RRR_0_OPCODE_X0 = 89, V1CMPLEU_RRR_0_OPCODE_X1 = 58, V1CMPLTSI_IMM8_OPCODE_X0 = 10, V1CMPLTSI_IMM8_OPCODE_X1 = 35, V1CMPLTS_RRR_0_OPCODE_X0 = 90, V1CMPLTS_RRR_0_OPCODE_X1 = 59, V1CMPLTUI_IMM8_OPCODE_X0 = 11, V1CMPLTUI_IMM8_OPCODE_X1 = 36, V1CMPLTU_RRR_0_OPCODE_X0 = 91, V1CMPLTU_RRR_0_OPCODE_X1 = 60, V1CMPNE_RRR_0_OPCODE_X0 = 92, V1CMPNE_RRR_0_OPCODE_X1 = 61, V1DDOTPUA_RRR_0_OPCODE_X0 = 161, V1DDOTPUSA_RRR_0_OPCODE_X0 = 93, V1DDOTPUS_RRR_0_OPCODE_X0 = 94, V1DDOTPU_RRR_0_OPCODE_X0 = 162, V1DOTPA_RRR_0_OPCODE_X0 = 95, V1DOTPUA_RRR_0_OPCODE_X0 = 163, V1DOTPUSA_RRR_0_OPCODE_X0 = 96, V1DOTPUS_RRR_0_OPCODE_X0 = 97, V1DOTPU_RRR_0_OPCODE_X0 = 164, V1DOTP_RRR_0_OPCODE_X0 = 98, V1INT_H_RRR_0_OPCODE_X0 = 99, V1INT_H_RRR_0_OPCODE_X1 = 62, V1INT_L_RRR_0_OPCODE_X0 = 100, V1INT_L_RRR_0_OPCODE_X1 = 63, V1MAXUI_IMM8_OPCODE_X0 = 12, V1MAXUI_IMM8_OPCODE_X1 = 37, V1MAXU_RRR_0_OPCODE_X0 = 101, V1MAXU_RRR_0_OPCODE_X1 = 64, V1MINUI_IMM8_OPCODE_X0 = 13, V1MINUI_IMM8_OPCODE_X1 = 38, V1MINU_RRR_0_OPCODE_X0 = 102, V1MINU_RRR_0_OPCODE_X1 = 65, V1MNZ_RRR_0_OPCODE_X0 = 103, V1MNZ_RRR_0_OPCODE_X1 = 66, V1MULTU_RRR_0_OPCODE_X0 = 104, V1MULUS_RRR_0_OPCODE_X0 = 105, V1MULU_RRR_0_OPCODE_X0 = 106, V1MZ_RRR_0_OPCODE_X0 = 107, V1MZ_RRR_0_OPCODE_X1 = 67, V1SADAU_RRR_0_OPCODE_X0 = 108, V1SADU_RRR_0_OPCODE_X0 = 109, V1SHLI_SHIFT_OPCODE_X0 = 7, V1SHLI_SHIFT_OPCODE_X1 = 7, V1SHL_RRR_0_OPCODE_X0 = 110, V1SHL_RRR_0_OPCODE_X1 = 68, V1SHRSI_SHIFT_OPCODE_X0 = 8, V1SHRSI_SHIFT_OPCODE_X1 = 8, V1SHRS_RRR_0_OPCODE_X0 = 111, V1SHRS_RRR_0_OPCODE_X1 = 69, V1SHRUI_SHIFT_OPCODE_X0 = 9, V1SHRUI_SHIFT_OPCODE_X1 = 9, V1SHRU_RRR_0_OPCODE_X0 = 112, V1SHRU_RRR_0_OPCODE_X1 = 70, V1SUBUC_RRR_0_OPCODE_X0 = 113, V1SUBUC_RRR_0_OPCODE_X1 = 71, V1SUB_RRR_0_OPCODE_X0 = 114, V1SUB_RRR_0_OPCODE_X1 = 72, V2ADDI_IMM8_OPCODE_X0 = 14, V2ADDI_IMM8_OPCODE_X1 = 39, V2ADDSC_RRR_0_OPCODE_X0 = 115, V2ADDSC_RRR_0_OPCODE_X1 = 73, V2ADD_RRR_0_OPCODE_X0 = 116, V2ADD_RRR_0_OPCODE_X1 = 74, V2ADIFFS_RRR_0_OPCODE_X0 = 117, V2AVGS_RRR_0_OPCODE_X0 = 118, V2CMPEQI_IMM8_OPCODE_X0 = 15, V2CMPEQI_IMM8_OPCODE_X1 = 40, V2CMPEQ_RRR_0_OPCODE_X0 = 119, V2CMPEQ_RRR_0_OPCODE_X1 = 75, V2CMPLES_RRR_0_OPCODE_X0 = 120, V2CMPLES_RRR_0_OPCODE_X1 = 76, V2CMPLEU_RRR_0_OPCODE_X0 = 121, V2CMPLEU_RRR_0_OPCODE_X1 = 77, V2CMPLTSI_IMM8_OPCODE_X0 = 16, V2CMPLTSI_IMM8_OPCODE_X1 = 41, V2CMPLTS_RRR_0_OPCODE_X0 = 122, V2CMPLTS_RRR_0_OPCODE_X1 = 78, V2CMPLTUI_IMM8_OPCODE_X0 = 17, V2CMPLTUI_IMM8_OPCODE_X1 = 42, V2CMPLTU_RRR_0_OPCODE_X0 = 123, V2CMPLTU_RRR_0_OPCODE_X1 = 79, V2CMPNE_RRR_0_OPCODE_X0 = 124, V2CMPNE_RRR_0_OPCODE_X1 = 80, V2DOTPA_RRR_0_OPCODE_X0 = 125, V2DOTP_RRR_0_OPCODE_X0 = 126, V2INT_H_RRR_0_OPCODE_X0 = 127, V2INT_H_RRR_0_OPCODE_X1 = 81, V2INT_L_RRR_0_OPCODE_X0 = 128, V2INT_L_RRR_0_OPCODE_X1 = 82, V2MAXSI_IMM8_OPCODE_X0 = 18, V2MAXSI_IMM8_OPCODE_X1 = 43, V2MAXS_RRR_0_OPCODE_X0 = 129, V2MAXS_RRR_0_OPCODE_X1 = 83, V2MINSI_IMM8_OPCODE_X0 = 19, V2MINSI_IMM8_OPCODE_X1 = 44, V2MINS_RRR_0_OPCODE_X0 = 130, V2MINS_RRR_0_OPCODE_X1 = 84, V2MNZ_RRR_0_OPCODE_X0 = 131, V2MNZ_RRR_0_OPCODE_X1 = 85, V2MULFSC_RRR_0_OPCODE_X0 = 132, V2MULS_RRR_0_OPCODE_X0 = 133, V2MULTS_RRR_0_OPCODE_X0 = 134, V2MZ_RRR_0_OPCODE_X0 = 135, V2MZ_RRR_0_OPCODE_X1 = 86, V2PACKH_RRR_0_OPCODE_X0 = 136, V2PACKH_RRR_0_OPCODE_X1 = 87, V2PACKL_RRR_0_OPCODE_X0 = 137, V2PACKL_RRR_0_OPCODE_X1 = 88, V2PACKUC_RRR_0_OPCODE_X0 = 138, V2PACKUC_RRR_0_OPCODE_X1 = 89, V2SADAS_RRR_0_OPCODE_X0 = 139, V2SADAU_RRR_0_OPCODE_X0 = 140, V2SADS_RRR_0_OPCODE_X0 = 141, V2SADU_RRR_0_OPCODE_X0 = 142, V2SHLI_SHIFT_OPCODE_X0 = 10, V2SHLI_SHIFT_OPCODE_X1 = 10, V2SHLSC_RRR_0_OPCODE_X0 = 143, V2SHLSC_RRR_0_OPCODE_X1 = 90, V2SHL_RRR_0_OPCODE_X0 = 144, V2SHL_RRR_0_OPCODE_X1 = 91, V2SHRSI_SHIFT_OPCODE_X0 = 11, V2SHRSI_SHIFT_OPCODE_X1 = 11, V2SHRS_RRR_0_OPCODE_X0 = 145, V2SHRS_RRR_0_OPCODE_X1 = 92, V2SHRUI_SHIFT_OPCODE_X0 = 12, V2SHRUI_SHIFT_OPCODE_X1 = 12, V2SHRU_RRR_0_OPCODE_X0 = 146, V2SHRU_RRR_0_OPCODE_X1 = 93, V2SUBSC_RRR_0_OPCODE_X0 = 147, V2SUBSC_RRR_0_OPCODE_X1 = 94, V2SUB_RRR_0_OPCODE_X0 = 148, V2SUB_RRR_0_OPCODE_X1 = 95, V4ADDSC_RRR_0_OPCODE_X0 = 149, V4ADDSC_RRR_0_OPCODE_X1 = 96, V4ADD_RRR_0_OPCODE_X0 = 150, V4ADD_RRR_0_OPCODE_X1 = 97, V4INT_H_RRR_0_OPCODE_X0 = 151, V4INT_H_RRR_0_OPCODE_X1 = 98, V4INT_L_RRR_0_OPCODE_X0 = 152, V4INT_L_RRR_0_OPCODE_X1 = 99, V4PACKSC_RRR_0_OPCODE_X0 = 153, V4PACKSC_RRR_0_OPCODE_X1 = 100, V4SHLSC_RRR_0_OPCODE_X0 = 154, V4SHLSC_RRR_0_OPCODE_X1 = 101, V4SHL_RRR_0_OPCODE_X0 = 155, V4SHL_RRR_0_OPCODE_X1 = 102, V4SHRS_RRR_0_OPCODE_X0 = 156, V4SHRS_RRR_0_OPCODE_X1 = 103, V4SHRU_RRR_0_OPCODE_X0 = 157, V4SHRU_RRR_0_OPCODE_X1 = 104, V4SUBSC_RRR_0_OPCODE_X0 = 158, V4SUBSC_RRR_0_OPCODE_X1 = 105, V4SUB_RRR_0_OPCODE_X0 = 159, V4SUB_RRR_0_OPCODE_X1 = 106, WH64_UNARY_OPCODE_X1 = 38, XORI_IMM8_OPCODE_X0 = 20, XORI_IMM8_OPCODE_X1 = 45, XOR_RRR_0_OPCODE_X0 = 160, XOR_RRR_0_OPCODE_X1 = 107, XOR_RRR_5_OPCODE_Y0 = 3, XOR_RRR_5_OPCODE_Y1 = 3 }; #endif /* __ASSEMBLER__ */ #endif /* __ARCH_OPCODE_H__ */