We've currently got 18 architectures in QEMU, and thus 18 target-xxx folders in the root folder of the QEMU source tree. More architectures (e.g. RISC-V, AVR) are likely to be included soon, too, so the main folder of the QEMU sources slowly gets quite overcrowded with the target-xxx folders. To disburden the main folder a little bit, let's move the target-xxx folders into a dedicated target/ folder, so that target-xxx/ simply becomes target/xxx/ instead. Acked-by: Laurent Vivier <laurent@vivier.eu> [m68k part] Acked-by: Bastian Koppelmann <kbastian@mail.uni-paderborn.de> [tricore part] Acked-by: Michael Walle <michael@walle.cc> [lm32 part] Acked-by: Cornelia Huck <cornelia.huck@de.ibm.com> [s390x part] Reviewed-by: Christian Borntraeger <borntraeger@de.ibm.com> [s390x part] Acked-by: Eduardo Habkost <ehabkost@redhat.com> [i386 part] Acked-by: Artyom Tarasenko <atar4qemu@gmail.com> [sparc part] Acked-by: Richard Henderson <rth@twiddle.net> [alpha part] Acked-by: Max Filippov <jcmvbkbc@gmail.com> [xtensa part] Reviewed-by: David Gibson <david@gibson.dropbear.id.au> [ppc part] Acked-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com> [crisµblaze part] Acked-by: Guan Xuetao <gxt@mprc.pku.edu.cn> [unicore32 part] Signed-off-by: Thomas Huth <thuth@redhat.com>
		
			
				
	
	
		
			1407 lines
		
	
	
		
			35 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1407 lines
		
	
	
		
			35 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* TILE-Gx opcode information.
 | |
|  *
 | |
|  * Copyright 2011 Tilera Corporation. All Rights Reserved.
 | |
|  *
 | |
|  *   This program is free software; you can redistribute it and/or
 | |
|  *   modify it under the terms of the GNU General Public License
 | |
|  *   as published by the Free Software Foundation, version 2.
 | |
|  *
 | |
|  *   This program is distributed in the hope that it will be useful, but
 | |
|  *   WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
 | |
|  *   NON INFRINGEMENT.  See the GNU General Public License for
 | |
|  *   more details.
 | |
|  *
 | |
|  *
 | |
|  *
 | |
|  *
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #ifndef OPCODE_TILEGX_H
 | |
| #define OPCODE_TILEGX_H
 | |
| 
 | |
| #ifndef __ASSEMBLER__
 | |
| 
 | |
| typedef uint64_t tilegx_bundle_bits;
 | |
| 
 | |
| /* These are the bits that determine if a bundle is in the X encoding. */
 | |
| #define TILEGX_BUNDLE_MODE_MASK ((tilegx_bundle_bits)3 << 62)
 | |
| 
 | |
| enum
 | |
| {
 | |
|   /* Maximum number of instructions in a bundle (2 for X, 3 for Y). */
 | |
|   TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE = 3,
 | |
| 
 | |
|   /* How many different pipeline encodings are there? X0, X1, Y0, Y1, Y2. */
 | |
|   TILEGX_NUM_PIPELINE_ENCODINGS = 5,
 | |
| 
 | |
|   /* Log base 2 of TILEGX_BUNDLE_SIZE_IN_BYTES. */
 | |
|   TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES = 3,
 | |
| 
 | |
|   /* Instructions take this many bytes. */
 | |
|   TILEGX_BUNDLE_SIZE_IN_BYTES = 1 << TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES,
 | |
| 
 | |
|   /* Log base 2 of TILEGX_BUNDLE_ALIGNMENT_IN_BYTES. */
 | |
|   TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES = 3,
 | |
| 
 | |
|   /* Bundles should be aligned modulo this number of bytes. */
 | |
|   TILEGX_BUNDLE_ALIGNMENT_IN_BYTES =
 | |
|     (1 << TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES),
 | |
| 
 | |
|   /* Number of registers (some are magic, such as network I/O). */
 | |
|   TILEGX_NUM_REGISTERS = 64,
 | |
| };
 | |
| 
 | |
| /* Make a few "tile_" variables to simplify common code between
 | |
|    architectures.  */
 | |
| 
 | |
| typedef tilegx_bundle_bits tile_bundle_bits;
 | |
| #define TILE_BUNDLE_SIZE_IN_BYTES TILEGX_BUNDLE_SIZE_IN_BYTES
 | |
| #define TILE_BUNDLE_ALIGNMENT_IN_BYTES TILEGX_BUNDLE_ALIGNMENT_IN_BYTES
 | |
| #define TILE_LOG2_BUNDLE_ALIGNMENT_IN_BYTES \
 | |
|   TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES
 | |
| #define TILE_BPT_BUNDLE TILEGX_BPT_BUNDLE
 | |
| 
 | |
| /* 64-bit pattern for a { bpt ; nop } bundle. */
 | |
| #define TILEGX_BPT_BUNDLE 0x286a44ae51485000ULL
 | |
| 
 | |
| static inline unsigned int
 | |
| get_BFEnd_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_BFOpcodeExtension_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 24)) & 0xf);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_BFStart_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 18)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_BrOff_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x0000003f) |
 | |
|          (((unsigned int)(n >> 37)) & 0x0001ffc0);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_BrType_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 54)) & 0x1f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Dest_Imm8_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x0000003f) |
 | |
|          (((unsigned int)(n >> 43)) & 0x000000c0);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Dest_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 0)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Dest_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Dest_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 0)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Dest_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm16_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0xffff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm16_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0xffff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8OpcodeExtension_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 20)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8OpcodeExtension_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 51)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Imm8_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0xff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_JumpOff_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x7ffffff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_JumpOpcodeExtension_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 58)) & 0x1);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_MF_Imm14_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 37)) & 0x3fff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_MT_Imm14_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 31)) & 0x0000003f) |
 | |
|          (((unsigned int)(n >> 37)) & 0x00003fc0);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Mode(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 62)) & 0x3);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Opcode_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 28)) & 0x7);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Opcode_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 59)) & 0x7);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Opcode_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 27)) & 0xf);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Opcode_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 58)) & 0xf);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_Opcode_Y2(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((n >> 26)) & 0x00000001) |
 | |
|          (((unsigned int)(n >> 56)) & 0x00000002);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_RRROpcodeExtension_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 18)) & 0x3ff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_RRROpcodeExtension_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 49)) & 0x3ff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_RRROpcodeExtension_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 18)) & 0x3);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_RRROpcodeExtension_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 49)) & 0x3);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShAmt_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShAmt_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShAmt_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShAmt_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShiftOpcodeExtension_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 18)) & 0x3ff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShiftOpcodeExtension_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 49)) & 0x3ff);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShiftOpcodeExtension_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 18)) & 0x3);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_ShiftOpcodeExtension_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 49)) & 0x3);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcA_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 6)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcA_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 37)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcA_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 6)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcA_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 37)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcA_Y2(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 20)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcBDest_Y2(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 51)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcB_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcB_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcB_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_SrcB_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_UnaryOpcodeExtension_X0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_UnaryOpcodeExtension_X1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_UnaryOpcodeExtension_Y0(tilegx_bundle_bits num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((n >> 12)) & 0x3f);
 | |
| }
 | |
| 
 | |
| static inline unsigned int
 | |
| get_UnaryOpcodeExtension_Y1(tilegx_bundle_bits n)
 | |
| {
 | |
|   return (((unsigned int)(n >> 43)) & 0x3f);
 | |
| }
 | |
| 
 | |
| 
 | |
| static inline int
 | |
| sign_extend(int n, int num_bits)
 | |
| {
 | |
|   int shift = (int)(sizeof(int) * 8 - num_bits);
 | |
|   return (n << shift) >> shift;
 | |
| }
 | |
| 
 | |
| 
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_BFEnd_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_BFOpcodeExtension_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xf) << 24);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_BFStart_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 18);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_BrOff_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
 | |
|          (((tilegx_bundle_bits)(n & 0x0001ffc0)) << 37);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_BrType_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x1f)) << 54);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Dest_Imm8_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
 | |
|          (((tilegx_bundle_bits)(n & 0x000000c0)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Dest_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 0);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Dest_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Dest_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 0);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Dest_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm16_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xffff) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm16_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0xffff)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8OpcodeExtension_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xff) << 20);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8OpcodeExtension_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0xff)) << 51);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xff) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0xff)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xff) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Imm8_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0xff)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_JumpOff_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x7ffffff)) << 31);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_JumpOpcodeExtension_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x1)) << 58);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_MF_Imm14_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3fff)) << 37);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_MT_Imm14_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
 | |
|          (((tilegx_bundle_bits)(n & 0x00003fc0)) << 37);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Mode(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3)) << 62);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Opcode_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x7) << 28);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Opcode_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x7)) << 59);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Opcode_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0xf) << 27);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Opcode_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0xf)) << 58);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_Opcode_Y2(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x00000001) << 26) |
 | |
|          (((tilegx_bundle_bits)(n & 0x00000002)) << 56);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_RRROpcodeExtension_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3ff) << 18);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_RRROpcodeExtension_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_RRROpcodeExtension_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3) << 18);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_RRROpcodeExtension_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3)) << 49);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShAmt_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShAmt_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShAmt_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShAmt_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShiftOpcodeExtension_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3ff) << 18);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShiftOpcodeExtension_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShiftOpcodeExtension_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3) << 18);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_ShiftOpcodeExtension_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3)) << 49);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcA_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 6);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcA_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcA_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 6);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcA_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcA_Y2(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 20);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcBDest_Y2(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 51);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcB_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcB_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcB_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_SrcB_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_UnaryOpcodeExtension_X0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_UnaryOpcodeExtension_X1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_UnaryOpcodeExtension_Y0(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return ((n & 0x3f) << 12);
 | |
| }
 | |
| 
 | |
| static inline tilegx_bundle_bits
 | |
| create_UnaryOpcodeExtension_Y1(int num)
 | |
| {
 | |
|   const unsigned int n = (unsigned int)num;
 | |
|   return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
 | |
| }
 | |
| 
 | |
| 
 | |
| enum
 | |
| {
 | |
|   ADDI_IMM8_OPCODE_X0 = 1,
 | |
|   ADDI_IMM8_OPCODE_X1 = 1,
 | |
|   ADDI_OPCODE_Y0 = 0,
 | |
|   ADDI_OPCODE_Y1 = 1,
 | |
|   ADDLI_OPCODE_X0 = 1,
 | |
|   ADDLI_OPCODE_X1 = 0,
 | |
|   ADDXI_IMM8_OPCODE_X0 = 2,
 | |
|   ADDXI_IMM8_OPCODE_X1 = 2,
 | |
|   ADDXI_OPCODE_Y0 = 1,
 | |
|   ADDXI_OPCODE_Y1 = 2,
 | |
|   ADDXLI_OPCODE_X0 = 2,
 | |
|   ADDXLI_OPCODE_X1 = 1,
 | |
|   ADDXSC_RRR_0_OPCODE_X0 = 1,
 | |
|   ADDXSC_RRR_0_OPCODE_X1 = 1,
 | |
|   ADDX_RRR_0_OPCODE_X0 = 2,
 | |
|   ADDX_RRR_0_OPCODE_X1 = 2,
 | |
|   ADDX_RRR_0_OPCODE_Y0 = 0,
 | |
|   ADDX_RRR_0_OPCODE_Y1 = 0,
 | |
|   ADD_RRR_0_OPCODE_X0 = 3,
 | |
|   ADD_RRR_0_OPCODE_X1 = 3,
 | |
|   ADD_RRR_0_OPCODE_Y0 = 1,
 | |
|   ADD_RRR_0_OPCODE_Y1 = 1,
 | |
|   ANDI_IMM8_OPCODE_X0 = 3,
 | |
|   ANDI_IMM8_OPCODE_X1 = 3,
 | |
|   ANDI_OPCODE_Y0 = 2,
 | |
|   ANDI_OPCODE_Y1 = 3,
 | |
|   AND_RRR_0_OPCODE_X0 = 4,
 | |
|   AND_RRR_0_OPCODE_X1 = 4,
 | |
|   AND_RRR_5_OPCODE_Y0 = 0,
 | |
|   AND_RRR_5_OPCODE_Y1 = 0,
 | |
|   BEQZT_BRANCH_OPCODE_X1 = 16,
 | |
|   BEQZ_BRANCH_OPCODE_X1 = 17,
 | |
|   BFEXTS_BF_OPCODE_X0 = 4,
 | |
|   BFEXTU_BF_OPCODE_X0 = 5,
 | |
|   BFINS_BF_OPCODE_X0 = 6,
 | |
|   BF_OPCODE_X0 = 3,
 | |
|   BGEZT_BRANCH_OPCODE_X1 = 18,
 | |
|   BGEZ_BRANCH_OPCODE_X1 = 19,
 | |
|   BGTZT_BRANCH_OPCODE_X1 = 20,
 | |
|   BGTZ_BRANCH_OPCODE_X1 = 21,
 | |
|   BLBCT_BRANCH_OPCODE_X1 = 22,
 | |
|   BLBC_BRANCH_OPCODE_X1 = 23,
 | |
|   BLBST_BRANCH_OPCODE_X1 = 24,
 | |
|   BLBS_BRANCH_OPCODE_X1 = 25,
 | |
|   BLEZT_BRANCH_OPCODE_X1 = 26,
 | |
|   BLEZ_BRANCH_OPCODE_X1 = 27,
 | |
|   BLTZT_BRANCH_OPCODE_X1 = 28,
 | |
|   BLTZ_BRANCH_OPCODE_X1 = 29,
 | |
|   BNEZT_BRANCH_OPCODE_X1 = 30,
 | |
|   BNEZ_BRANCH_OPCODE_X1 = 31,
 | |
|   BRANCH_OPCODE_X1 = 2,
 | |
|   CMOVEQZ_RRR_0_OPCODE_X0 = 5,
 | |
|   CMOVEQZ_RRR_4_OPCODE_Y0 = 0,
 | |
|   CMOVNEZ_RRR_0_OPCODE_X0 = 6,
 | |
|   CMOVNEZ_RRR_4_OPCODE_Y0 = 1,
 | |
|   CMPEQI_IMM8_OPCODE_X0 = 4,
 | |
|   CMPEQI_IMM8_OPCODE_X1 = 4,
 | |
|   CMPEQI_OPCODE_Y0 = 3,
 | |
|   CMPEQI_OPCODE_Y1 = 4,
 | |
|   CMPEQ_RRR_0_OPCODE_X0 = 7,
 | |
|   CMPEQ_RRR_0_OPCODE_X1 = 5,
 | |
|   CMPEQ_RRR_3_OPCODE_Y0 = 0,
 | |
|   CMPEQ_RRR_3_OPCODE_Y1 = 2,
 | |
|   CMPEXCH4_RRR_0_OPCODE_X1 = 6,
 | |
|   CMPEXCH_RRR_0_OPCODE_X1 = 7,
 | |
|   CMPLES_RRR_0_OPCODE_X0 = 8,
 | |
|   CMPLES_RRR_0_OPCODE_X1 = 8,
 | |
|   CMPLES_RRR_2_OPCODE_Y0 = 0,
 | |
|   CMPLES_RRR_2_OPCODE_Y1 = 0,
 | |
|   CMPLEU_RRR_0_OPCODE_X0 = 9,
 | |
|   CMPLEU_RRR_0_OPCODE_X1 = 9,
 | |
|   CMPLEU_RRR_2_OPCODE_Y0 = 1,
 | |
|   CMPLEU_RRR_2_OPCODE_Y1 = 1,
 | |
|   CMPLTSI_IMM8_OPCODE_X0 = 5,
 | |
|   CMPLTSI_IMM8_OPCODE_X1 = 5,
 | |
|   CMPLTSI_OPCODE_Y0 = 4,
 | |
|   CMPLTSI_OPCODE_Y1 = 5,
 | |
|   CMPLTS_RRR_0_OPCODE_X0 = 10,
 | |
|   CMPLTS_RRR_0_OPCODE_X1 = 10,
 | |
|   CMPLTS_RRR_2_OPCODE_Y0 = 2,
 | |
|   CMPLTS_RRR_2_OPCODE_Y1 = 2,
 | |
|   CMPLTUI_IMM8_OPCODE_X0 = 6,
 | |
|   CMPLTUI_IMM8_OPCODE_X1 = 6,
 | |
|   CMPLTU_RRR_0_OPCODE_X0 = 11,
 | |
|   CMPLTU_RRR_0_OPCODE_X1 = 11,
 | |
|   CMPLTU_RRR_2_OPCODE_Y0 = 3,
 | |
|   CMPLTU_RRR_2_OPCODE_Y1 = 3,
 | |
|   CMPNE_RRR_0_OPCODE_X0 = 12,
 | |
|   CMPNE_RRR_0_OPCODE_X1 = 12,
 | |
|   CMPNE_RRR_3_OPCODE_Y0 = 1,
 | |
|   CMPNE_RRR_3_OPCODE_Y1 = 3,
 | |
|   CMULAF_RRR_0_OPCODE_X0 = 13,
 | |
|   CMULA_RRR_0_OPCODE_X0 = 14,
 | |
|   CMULFR_RRR_0_OPCODE_X0 = 15,
 | |
|   CMULF_RRR_0_OPCODE_X0 = 16,
 | |
|   CMULHR_RRR_0_OPCODE_X0 = 17,
 | |
|   CMULH_RRR_0_OPCODE_X0 = 18,
 | |
|   CMUL_RRR_0_OPCODE_X0 = 19,
 | |
|   CNTLZ_UNARY_OPCODE_X0 = 1,
 | |
|   CNTLZ_UNARY_OPCODE_Y0 = 1,
 | |
|   CNTTZ_UNARY_OPCODE_X0 = 2,
 | |
|   CNTTZ_UNARY_OPCODE_Y0 = 2,
 | |
|   CRC32_32_RRR_0_OPCODE_X0 = 20,
 | |
|   CRC32_8_RRR_0_OPCODE_X0 = 21,
 | |
|   DBLALIGN2_RRR_0_OPCODE_X0 = 22,
 | |
|   DBLALIGN2_RRR_0_OPCODE_X1 = 13,
 | |
|   DBLALIGN4_RRR_0_OPCODE_X0 = 23,
 | |
|   DBLALIGN4_RRR_0_OPCODE_X1 = 14,
 | |
|   DBLALIGN6_RRR_0_OPCODE_X0 = 24,
 | |
|   DBLALIGN6_RRR_0_OPCODE_X1 = 15,
 | |
|   DBLALIGN_RRR_0_OPCODE_X0 = 25,
 | |
|   DRAIN_UNARY_OPCODE_X1 = 1,
 | |
|   DTLBPR_UNARY_OPCODE_X1 = 2,
 | |
|   EXCH4_RRR_0_OPCODE_X1 = 16,
 | |
|   EXCH_RRR_0_OPCODE_X1 = 17,
 | |
|   FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 = 26,
 | |
|   FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 = 27,
 | |
|   FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 = 28,
 | |
|   FDOUBLE_PACK1_RRR_0_OPCODE_X0 = 29,
 | |
|   FDOUBLE_PACK2_RRR_0_OPCODE_X0 = 30,
 | |
|   FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 = 31,
 | |
|   FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 = 32,
 | |
|   FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 = 33,
 | |
|   FETCHADD4_RRR_0_OPCODE_X1 = 18,
 | |
|   FETCHADDGEZ4_RRR_0_OPCODE_X1 = 19,
 | |
|   FETCHADDGEZ_RRR_0_OPCODE_X1 = 20,
 | |
|   FETCHADD_RRR_0_OPCODE_X1 = 21,
 | |
|   FETCHAND4_RRR_0_OPCODE_X1 = 22,
 | |
|   FETCHAND_RRR_0_OPCODE_X1 = 23,
 | |
|   FETCHOR4_RRR_0_OPCODE_X1 = 24,
 | |
|   FETCHOR_RRR_0_OPCODE_X1 = 25,
 | |
|   FINV_UNARY_OPCODE_X1 = 3,
 | |
|   FLUSHWB_UNARY_OPCODE_X1 = 4,
 | |
|   FLUSH_UNARY_OPCODE_X1 = 5,
 | |
|   FNOP_UNARY_OPCODE_X0 = 3,
 | |
|   FNOP_UNARY_OPCODE_X1 = 6,
 | |
|   FNOP_UNARY_OPCODE_Y0 = 3,
 | |
|   FNOP_UNARY_OPCODE_Y1 = 8,
 | |
|   FSINGLE_ADD1_RRR_0_OPCODE_X0 = 34,
 | |
|   FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 = 35,
 | |
|   FSINGLE_MUL1_RRR_0_OPCODE_X0 = 36,
 | |
|   FSINGLE_MUL2_RRR_0_OPCODE_X0 = 37,
 | |
|   FSINGLE_PACK1_UNARY_OPCODE_X0 = 4,
 | |
|   FSINGLE_PACK1_UNARY_OPCODE_Y0 = 4,
 | |
|   FSINGLE_PACK2_RRR_0_OPCODE_X0 = 38,
 | |
|   FSINGLE_SUB1_RRR_0_OPCODE_X0 = 39,
 | |
|   ICOH_UNARY_OPCODE_X1 = 7,
 | |
|   ILL_UNARY_OPCODE_X1 = 8,
 | |
|   ILL_UNARY_OPCODE_Y1 = 9,
 | |
|   IMM8_OPCODE_X0 = 4,
 | |
|   IMM8_OPCODE_X1 = 3,
 | |
|   INV_UNARY_OPCODE_X1 = 9,
 | |
|   IRET_UNARY_OPCODE_X1 = 10,
 | |
|   JALRP_UNARY_OPCODE_X1 = 11,
 | |
|   JALRP_UNARY_OPCODE_Y1 = 10,
 | |
|   JALR_UNARY_OPCODE_X1 = 12,
 | |
|   JALR_UNARY_OPCODE_Y1 = 11,
 | |
|   JAL_JUMP_OPCODE_X1 = 0,
 | |
|   JRP_UNARY_OPCODE_X1 = 13,
 | |
|   JRP_UNARY_OPCODE_Y1 = 12,
 | |
|   JR_UNARY_OPCODE_X1 = 14,
 | |
|   JR_UNARY_OPCODE_Y1 = 13,
 | |
|   JUMP_OPCODE_X1 = 4,
 | |
|   J_JUMP_OPCODE_X1 = 1,
 | |
|   LD1S_ADD_IMM8_OPCODE_X1 = 7,
 | |
|   LD1S_OPCODE_Y2 = 0,
 | |
|   LD1S_UNARY_OPCODE_X1 = 15,
 | |
|   LD1U_ADD_IMM8_OPCODE_X1 = 8,
 | |
|   LD1U_OPCODE_Y2 = 1,
 | |
|   LD1U_UNARY_OPCODE_X1 = 16,
 | |
|   LD2S_ADD_IMM8_OPCODE_X1 = 9,
 | |
|   LD2S_OPCODE_Y2 = 2,
 | |
|   LD2S_UNARY_OPCODE_X1 = 17,
 | |
|   LD2U_ADD_IMM8_OPCODE_X1 = 10,
 | |
|   LD2U_OPCODE_Y2 = 3,
 | |
|   LD2U_UNARY_OPCODE_X1 = 18,
 | |
|   LD4S_ADD_IMM8_OPCODE_X1 = 11,
 | |
|   LD4S_OPCODE_Y2 = 1,
 | |
|   LD4S_UNARY_OPCODE_X1 = 19,
 | |
|   LD4U_ADD_IMM8_OPCODE_X1 = 12,
 | |
|   LD4U_OPCODE_Y2 = 2,
 | |
|   LD4U_UNARY_OPCODE_X1 = 20,
 | |
|   LDNA_UNARY_OPCODE_X1 = 21,
 | |
|   LDNT1S_ADD_IMM8_OPCODE_X1 = 13,
 | |
|   LDNT1S_UNARY_OPCODE_X1 = 22,
 | |
|   LDNT1U_ADD_IMM8_OPCODE_X1 = 14,
 | |
|   LDNT1U_UNARY_OPCODE_X1 = 23,
 | |
|   LDNT2S_ADD_IMM8_OPCODE_X1 = 15,
 | |
|   LDNT2S_UNARY_OPCODE_X1 = 24,
 | |
|   LDNT2U_ADD_IMM8_OPCODE_X1 = 16,
 | |
|   LDNT2U_UNARY_OPCODE_X1 = 25,
 | |
|   LDNT4S_ADD_IMM8_OPCODE_X1 = 17,
 | |
|   LDNT4S_UNARY_OPCODE_X1 = 26,
 | |
|   LDNT4U_ADD_IMM8_OPCODE_X1 = 18,
 | |
|   LDNT4U_UNARY_OPCODE_X1 = 27,
 | |
|   LDNT_ADD_IMM8_OPCODE_X1 = 19,
 | |
|   LDNT_UNARY_OPCODE_X1 = 28,
 | |
|   LD_ADD_IMM8_OPCODE_X1 = 20,
 | |
|   LD_OPCODE_Y2 = 3,
 | |
|   LD_UNARY_OPCODE_X1 = 29,
 | |
|   LNK_UNARY_OPCODE_X1 = 30,
 | |
|   LNK_UNARY_OPCODE_Y1 = 14,
 | |
|   LDNA_ADD_IMM8_OPCODE_X1 = 21,
 | |
|   MFSPR_IMM8_OPCODE_X1 = 22,
 | |
|   MF_UNARY_OPCODE_X1 = 31,
 | |
|   MM_BF_OPCODE_X0 = 7,
 | |
|   MNZ_RRR_0_OPCODE_X0 = 40,
 | |
|   MNZ_RRR_0_OPCODE_X1 = 26,
 | |
|   MNZ_RRR_4_OPCODE_Y0 = 2,
 | |
|   MNZ_RRR_4_OPCODE_Y1 = 2,
 | |
|   MODE_OPCODE_YA2 = 1,
 | |
|   MODE_OPCODE_YB2 = 2,
 | |
|   MODE_OPCODE_YC2 = 3,
 | |
|   MTSPR_IMM8_OPCODE_X1 = 23,
 | |
|   MULAX_RRR_0_OPCODE_X0 = 41,
 | |
|   MULAX_RRR_3_OPCODE_Y0 = 2,
 | |
|   MULA_HS_HS_RRR_0_OPCODE_X0 = 42,
 | |
|   MULA_HS_HS_RRR_9_OPCODE_Y0 = 0,
 | |
|   MULA_HS_HU_RRR_0_OPCODE_X0 = 43,
 | |
|   MULA_HS_LS_RRR_0_OPCODE_X0 = 44,
 | |
|   MULA_HS_LU_RRR_0_OPCODE_X0 = 45,
 | |
|   MULA_HU_HU_RRR_0_OPCODE_X0 = 46,
 | |
|   MULA_HU_HU_RRR_9_OPCODE_Y0 = 1,
 | |
|   MULA_HU_LS_RRR_0_OPCODE_X0 = 47,
 | |
|   MULA_HU_LU_RRR_0_OPCODE_X0 = 48,
 | |
|   MULA_LS_LS_RRR_0_OPCODE_X0 = 49,
 | |
|   MULA_LS_LS_RRR_9_OPCODE_Y0 = 2,
 | |
|   MULA_LS_LU_RRR_0_OPCODE_X0 = 50,
 | |
|   MULA_LU_LU_RRR_0_OPCODE_X0 = 51,
 | |
|   MULA_LU_LU_RRR_9_OPCODE_Y0 = 3,
 | |
|   MULX_RRR_0_OPCODE_X0 = 52,
 | |
|   MULX_RRR_3_OPCODE_Y0 = 3,
 | |
|   MUL_HS_HS_RRR_0_OPCODE_X0 = 53,
 | |
|   MUL_HS_HS_RRR_8_OPCODE_Y0 = 0,
 | |
|   MUL_HS_HU_RRR_0_OPCODE_X0 = 54,
 | |
|   MUL_HS_LS_RRR_0_OPCODE_X0 = 55,
 | |
|   MUL_HS_LU_RRR_0_OPCODE_X0 = 56,
 | |
|   MUL_HU_HU_RRR_0_OPCODE_X0 = 57,
 | |
|   MUL_HU_HU_RRR_8_OPCODE_Y0 = 1,
 | |
|   MUL_HU_LS_RRR_0_OPCODE_X0 = 58,
 | |
|   MUL_HU_LU_RRR_0_OPCODE_X0 = 59,
 | |
|   MUL_LS_LS_RRR_0_OPCODE_X0 = 60,
 | |
|   MUL_LS_LS_RRR_8_OPCODE_Y0 = 2,
 | |
|   MUL_LS_LU_RRR_0_OPCODE_X0 = 61,
 | |
|   MUL_LU_LU_RRR_0_OPCODE_X0 = 62,
 | |
|   MUL_LU_LU_RRR_8_OPCODE_Y0 = 3,
 | |
|   MZ_RRR_0_OPCODE_X0 = 63,
 | |
|   MZ_RRR_0_OPCODE_X1 = 27,
 | |
|   MZ_RRR_4_OPCODE_Y0 = 3,
 | |
|   MZ_RRR_4_OPCODE_Y1 = 3,
 | |
|   NAP_UNARY_OPCODE_X1 = 32,
 | |
|   NOP_UNARY_OPCODE_X0 = 5,
 | |
|   NOP_UNARY_OPCODE_X1 = 33,
 | |
|   NOP_UNARY_OPCODE_Y0 = 5,
 | |
|   NOP_UNARY_OPCODE_Y1 = 15,
 | |
|   NOR_RRR_0_OPCODE_X0 = 64,
 | |
|   NOR_RRR_0_OPCODE_X1 = 28,
 | |
|   NOR_RRR_5_OPCODE_Y0 = 1,
 | |
|   NOR_RRR_5_OPCODE_Y1 = 1,
 | |
|   ORI_IMM8_OPCODE_X0 = 7,
 | |
|   ORI_IMM8_OPCODE_X1 = 24,
 | |
|   OR_RRR_0_OPCODE_X0 = 65,
 | |
|   OR_RRR_0_OPCODE_X1 = 29,
 | |
|   OR_RRR_5_OPCODE_Y0 = 2,
 | |
|   OR_RRR_5_OPCODE_Y1 = 2,
 | |
|   PCNT_UNARY_OPCODE_X0 = 6,
 | |
|   PCNT_UNARY_OPCODE_Y0 = 6,
 | |
|   REVBITS_UNARY_OPCODE_X0 = 7,
 | |
|   REVBITS_UNARY_OPCODE_Y0 = 7,
 | |
|   REVBYTES_UNARY_OPCODE_X0 = 8,
 | |
|   REVBYTES_UNARY_OPCODE_Y0 = 8,
 | |
|   ROTLI_SHIFT_OPCODE_X0 = 1,
 | |
|   ROTLI_SHIFT_OPCODE_X1 = 1,
 | |
|   ROTLI_SHIFT_OPCODE_Y0 = 0,
 | |
|   ROTLI_SHIFT_OPCODE_Y1 = 0,
 | |
|   ROTL_RRR_0_OPCODE_X0 = 66,
 | |
|   ROTL_RRR_0_OPCODE_X1 = 30,
 | |
|   ROTL_RRR_6_OPCODE_Y0 = 0,
 | |
|   ROTL_RRR_6_OPCODE_Y1 = 0,
 | |
|   RRR_0_OPCODE_X0 = 5,
 | |
|   RRR_0_OPCODE_X1 = 5,
 | |
|   RRR_0_OPCODE_Y0 = 5,
 | |
|   RRR_0_OPCODE_Y1 = 6,
 | |
|   RRR_1_OPCODE_Y0 = 6,
 | |
|   RRR_1_OPCODE_Y1 = 7,
 | |
|   RRR_2_OPCODE_Y0 = 7,
 | |
|   RRR_2_OPCODE_Y1 = 8,
 | |
|   RRR_3_OPCODE_Y0 = 8,
 | |
|   RRR_3_OPCODE_Y1 = 9,
 | |
|   RRR_4_OPCODE_Y0 = 9,
 | |
|   RRR_4_OPCODE_Y1 = 10,
 | |
|   RRR_5_OPCODE_Y0 = 10,
 | |
|   RRR_5_OPCODE_Y1 = 11,
 | |
|   RRR_6_OPCODE_Y0 = 11,
 | |
|   RRR_6_OPCODE_Y1 = 12,
 | |
|   RRR_7_OPCODE_Y0 = 12,
 | |
|   RRR_7_OPCODE_Y1 = 13,
 | |
|   RRR_8_OPCODE_Y0 = 13,
 | |
|   RRR_9_OPCODE_Y0 = 14,
 | |
|   SHIFT_OPCODE_X0 = 6,
 | |
|   SHIFT_OPCODE_X1 = 6,
 | |
|   SHIFT_OPCODE_Y0 = 15,
 | |
|   SHIFT_OPCODE_Y1 = 14,
 | |
|   SHL16INSLI_OPCODE_X0 = 7,
 | |
|   SHL16INSLI_OPCODE_X1 = 7,
 | |
|   SHL1ADDX_RRR_0_OPCODE_X0 = 67,
 | |
|   SHL1ADDX_RRR_0_OPCODE_X1 = 31,
 | |
|   SHL1ADDX_RRR_7_OPCODE_Y0 = 1,
 | |
|   SHL1ADDX_RRR_7_OPCODE_Y1 = 1,
 | |
|   SHL1ADD_RRR_0_OPCODE_X0 = 68,
 | |
|   SHL1ADD_RRR_0_OPCODE_X1 = 32,
 | |
|   SHL1ADD_RRR_1_OPCODE_Y0 = 0,
 | |
|   SHL1ADD_RRR_1_OPCODE_Y1 = 0,
 | |
|   SHL2ADDX_RRR_0_OPCODE_X0 = 69,
 | |
|   SHL2ADDX_RRR_0_OPCODE_X1 = 33,
 | |
|   SHL2ADDX_RRR_7_OPCODE_Y0 = 2,
 | |
|   SHL2ADDX_RRR_7_OPCODE_Y1 = 2,
 | |
|   SHL2ADD_RRR_0_OPCODE_X0 = 70,
 | |
|   SHL2ADD_RRR_0_OPCODE_X1 = 34,
 | |
|   SHL2ADD_RRR_1_OPCODE_Y0 = 1,
 | |
|   SHL2ADD_RRR_1_OPCODE_Y1 = 1,
 | |
|   SHL3ADDX_RRR_0_OPCODE_X0 = 71,
 | |
|   SHL3ADDX_RRR_0_OPCODE_X1 = 35,
 | |
|   SHL3ADDX_RRR_7_OPCODE_Y0 = 3,
 | |
|   SHL3ADDX_RRR_7_OPCODE_Y1 = 3,
 | |
|   SHL3ADD_RRR_0_OPCODE_X0 = 72,
 | |
|   SHL3ADD_RRR_0_OPCODE_X1 = 36,
 | |
|   SHL3ADD_RRR_1_OPCODE_Y0 = 2,
 | |
|   SHL3ADD_RRR_1_OPCODE_Y1 = 2,
 | |
|   SHLI_SHIFT_OPCODE_X0 = 2,
 | |
|   SHLI_SHIFT_OPCODE_X1 = 2,
 | |
|   SHLI_SHIFT_OPCODE_Y0 = 1,
 | |
|   SHLI_SHIFT_OPCODE_Y1 = 1,
 | |
|   SHLXI_SHIFT_OPCODE_X0 = 3,
 | |
|   SHLXI_SHIFT_OPCODE_X1 = 3,
 | |
|   SHLX_RRR_0_OPCODE_X0 = 73,
 | |
|   SHLX_RRR_0_OPCODE_X1 = 37,
 | |
|   SHL_RRR_0_OPCODE_X0 = 74,
 | |
|   SHL_RRR_0_OPCODE_X1 = 38,
 | |
|   SHL_RRR_6_OPCODE_Y0 = 1,
 | |
|   SHL_RRR_6_OPCODE_Y1 = 1,
 | |
|   SHRSI_SHIFT_OPCODE_X0 = 4,
 | |
|   SHRSI_SHIFT_OPCODE_X1 = 4,
 | |
|   SHRSI_SHIFT_OPCODE_Y0 = 2,
 | |
|   SHRSI_SHIFT_OPCODE_Y1 = 2,
 | |
|   SHRS_RRR_0_OPCODE_X0 = 75,
 | |
|   SHRS_RRR_0_OPCODE_X1 = 39,
 | |
|   SHRS_RRR_6_OPCODE_Y0 = 2,
 | |
|   SHRS_RRR_6_OPCODE_Y1 = 2,
 | |
|   SHRUI_SHIFT_OPCODE_X0 = 5,
 | |
|   SHRUI_SHIFT_OPCODE_X1 = 5,
 | |
|   SHRUI_SHIFT_OPCODE_Y0 = 3,
 | |
|   SHRUI_SHIFT_OPCODE_Y1 = 3,
 | |
|   SHRUXI_SHIFT_OPCODE_X0 = 6,
 | |
|   SHRUXI_SHIFT_OPCODE_X1 = 6,
 | |
|   SHRUX_RRR_0_OPCODE_X0 = 76,
 | |
|   SHRUX_RRR_0_OPCODE_X1 = 40,
 | |
|   SHRU_RRR_0_OPCODE_X0 = 77,
 | |
|   SHRU_RRR_0_OPCODE_X1 = 41,
 | |
|   SHRU_RRR_6_OPCODE_Y0 = 3,
 | |
|   SHRU_RRR_6_OPCODE_Y1 = 3,
 | |
|   SHUFFLEBYTES_RRR_0_OPCODE_X0 = 78,
 | |
|   ST1_ADD_IMM8_OPCODE_X1 = 25,
 | |
|   ST1_OPCODE_Y2 = 0,
 | |
|   ST1_RRR_0_OPCODE_X1 = 42,
 | |
|   ST2_ADD_IMM8_OPCODE_X1 = 26,
 | |
|   ST2_OPCODE_Y2 = 1,
 | |
|   ST2_RRR_0_OPCODE_X1 = 43,
 | |
|   ST4_ADD_IMM8_OPCODE_X1 = 27,
 | |
|   ST4_OPCODE_Y2 = 2,
 | |
|   ST4_RRR_0_OPCODE_X1 = 44,
 | |
|   STNT1_ADD_IMM8_OPCODE_X1 = 28,
 | |
|   STNT1_RRR_0_OPCODE_X1 = 45,
 | |
|   STNT2_ADD_IMM8_OPCODE_X1 = 29,
 | |
|   STNT2_RRR_0_OPCODE_X1 = 46,
 | |
|   STNT4_ADD_IMM8_OPCODE_X1 = 30,
 | |
|   STNT4_RRR_0_OPCODE_X1 = 47,
 | |
|   STNT_ADD_IMM8_OPCODE_X1 = 31,
 | |
|   STNT_RRR_0_OPCODE_X1 = 48,
 | |
|   ST_ADD_IMM8_OPCODE_X1 = 32,
 | |
|   ST_OPCODE_Y2 = 3,
 | |
|   ST_RRR_0_OPCODE_X1 = 49,
 | |
|   SUBXSC_RRR_0_OPCODE_X0 = 79,
 | |
|   SUBXSC_RRR_0_OPCODE_X1 = 50,
 | |
|   SUBX_RRR_0_OPCODE_X0 = 80,
 | |
|   SUBX_RRR_0_OPCODE_X1 = 51,
 | |
|   SUBX_RRR_0_OPCODE_Y0 = 2,
 | |
|   SUBX_RRR_0_OPCODE_Y1 = 2,
 | |
|   SUB_RRR_0_OPCODE_X0 = 81,
 | |
|   SUB_RRR_0_OPCODE_X1 = 52,
 | |
|   SUB_RRR_0_OPCODE_Y0 = 3,
 | |
|   SUB_RRR_0_OPCODE_Y1 = 3,
 | |
|   SWINT0_UNARY_OPCODE_X1 = 34,
 | |
|   SWINT1_UNARY_OPCODE_X1 = 35,
 | |
|   SWINT2_UNARY_OPCODE_X1 = 36,
 | |
|   SWINT3_UNARY_OPCODE_X1 = 37,
 | |
|   TBLIDXB0_UNARY_OPCODE_X0 = 9,
 | |
|   TBLIDXB0_UNARY_OPCODE_Y0 = 9,
 | |
|   TBLIDXB1_UNARY_OPCODE_X0 = 10,
 | |
|   TBLIDXB1_UNARY_OPCODE_Y0 = 10,
 | |
|   TBLIDXB2_UNARY_OPCODE_X0 = 11,
 | |
|   TBLIDXB2_UNARY_OPCODE_Y0 = 11,
 | |
|   TBLIDXB3_UNARY_OPCODE_X0 = 12,
 | |
|   TBLIDXB3_UNARY_OPCODE_Y0 = 12,
 | |
|   UNARY_RRR_0_OPCODE_X0 = 82,
 | |
|   UNARY_RRR_0_OPCODE_X1 = 53,
 | |
|   UNARY_RRR_1_OPCODE_Y0 = 3,
 | |
|   UNARY_RRR_1_OPCODE_Y1 = 3,
 | |
|   V1ADDI_IMM8_OPCODE_X0 = 8,
 | |
|   V1ADDI_IMM8_OPCODE_X1 = 33,
 | |
|   V1ADDUC_RRR_0_OPCODE_X0 = 83,
 | |
|   V1ADDUC_RRR_0_OPCODE_X1 = 54,
 | |
|   V1ADD_RRR_0_OPCODE_X0 = 84,
 | |
|   V1ADD_RRR_0_OPCODE_X1 = 55,
 | |
|   V1ADIFFU_RRR_0_OPCODE_X0 = 85,
 | |
|   V1AVGU_RRR_0_OPCODE_X0 = 86,
 | |
|   V1CMPEQI_IMM8_OPCODE_X0 = 9,
 | |
|   V1CMPEQI_IMM8_OPCODE_X1 = 34,
 | |
|   V1CMPEQ_RRR_0_OPCODE_X0 = 87,
 | |
|   V1CMPEQ_RRR_0_OPCODE_X1 = 56,
 | |
|   V1CMPLES_RRR_0_OPCODE_X0 = 88,
 | |
|   V1CMPLES_RRR_0_OPCODE_X1 = 57,
 | |
|   V1CMPLEU_RRR_0_OPCODE_X0 = 89,
 | |
|   V1CMPLEU_RRR_0_OPCODE_X1 = 58,
 | |
|   V1CMPLTSI_IMM8_OPCODE_X0 = 10,
 | |
|   V1CMPLTSI_IMM8_OPCODE_X1 = 35,
 | |
|   V1CMPLTS_RRR_0_OPCODE_X0 = 90,
 | |
|   V1CMPLTS_RRR_0_OPCODE_X1 = 59,
 | |
|   V1CMPLTUI_IMM8_OPCODE_X0 = 11,
 | |
|   V1CMPLTUI_IMM8_OPCODE_X1 = 36,
 | |
|   V1CMPLTU_RRR_0_OPCODE_X0 = 91,
 | |
|   V1CMPLTU_RRR_0_OPCODE_X1 = 60,
 | |
|   V1CMPNE_RRR_0_OPCODE_X0 = 92,
 | |
|   V1CMPNE_RRR_0_OPCODE_X1 = 61,
 | |
|   V1DDOTPUA_RRR_0_OPCODE_X0 = 161,
 | |
|   V1DDOTPUSA_RRR_0_OPCODE_X0 = 93,
 | |
|   V1DDOTPUS_RRR_0_OPCODE_X0 = 94,
 | |
|   V1DDOTPU_RRR_0_OPCODE_X0 = 162,
 | |
|   V1DOTPA_RRR_0_OPCODE_X0 = 95,
 | |
|   V1DOTPUA_RRR_0_OPCODE_X0 = 163,
 | |
|   V1DOTPUSA_RRR_0_OPCODE_X0 = 96,
 | |
|   V1DOTPUS_RRR_0_OPCODE_X0 = 97,
 | |
|   V1DOTPU_RRR_0_OPCODE_X0 = 164,
 | |
|   V1DOTP_RRR_0_OPCODE_X0 = 98,
 | |
|   V1INT_H_RRR_0_OPCODE_X0 = 99,
 | |
|   V1INT_H_RRR_0_OPCODE_X1 = 62,
 | |
|   V1INT_L_RRR_0_OPCODE_X0 = 100,
 | |
|   V1INT_L_RRR_0_OPCODE_X1 = 63,
 | |
|   V1MAXUI_IMM8_OPCODE_X0 = 12,
 | |
|   V1MAXUI_IMM8_OPCODE_X1 = 37,
 | |
|   V1MAXU_RRR_0_OPCODE_X0 = 101,
 | |
|   V1MAXU_RRR_0_OPCODE_X1 = 64,
 | |
|   V1MINUI_IMM8_OPCODE_X0 = 13,
 | |
|   V1MINUI_IMM8_OPCODE_X1 = 38,
 | |
|   V1MINU_RRR_0_OPCODE_X0 = 102,
 | |
|   V1MINU_RRR_0_OPCODE_X1 = 65,
 | |
|   V1MNZ_RRR_0_OPCODE_X0 = 103,
 | |
|   V1MNZ_RRR_0_OPCODE_X1 = 66,
 | |
|   V1MULTU_RRR_0_OPCODE_X0 = 104,
 | |
|   V1MULUS_RRR_0_OPCODE_X0 = 105,
 | |
|   V1MULU_RRR_0_OPCODE_X0 = 106,
 | |
|   V1MZ_RRR_0_OPCODE_X0 = 107,
 | |
|   V1MZ_RRR_0_OPCODE_X1 = 67,
 | |
|   V1SADAU_RRR_0_OPCODE_X0 = 108,
 | |
|   V1SADU_RRR_0_OPCODE_X0 = 109,
 | |
|   V1SHLI_SHIFT_OPCODE_X0 = 7,
 | |
|   V1SHLI_SHIFT_OPCODE_X1 = 7,
 | |
|   V1SHL_RRR_0_OPCODE_X0 = 110,
 | |
|   V1SHL_RRR_0_OPCODE_X1 = 68,
 | |
|   V1SHRSI_SHIFT_OPCODE_X0 = 8,
 | |
|   V1SHRSI_SHIFT_OPCODE_X1 = 8,
 | |
|   V1SHRS_RRR_0_OPCODE_X0 = 111,
 | |
|   V1SHRS_RRR_0_OPCODE_X1 = 69,
 | |
|   V1SHRUI_SHIFT_OPCODE_X0 = 9,
 | |
|   V1SHRUI_SHIFT_OPCODE_X1 = 9,
 | |
|   V1SHRU_RRR_0_OPCODE_X0 = 112,
 | |
|   V1SHRU_RRR_0_OPCODE_X1 = 70,
 | |
|   V1SUBUC_RRR_0_OPCODE_X0 = 113,
 | |
|   V1SUBUC_RRR_0_OPCODE_X1 = 71,
 | |
|   V1SUB_RRR_0_OPCODE_X0 = 114,
 | |
|   V1SUB_RRR_0_OPCODE_X1 = 72,
 | |
|   V2ADDI_IMM8_OPCODE_X0 = 14,
 | |
|   V2ADDI_IMM8_OPCODE_X1 = 39,
 | |
|   V2ADDSC_RRR_0_OPCODE_X0 = 115,
 | |
|   V2ADDSC_RRR_0_OPCODE_X1 = 73,
 | |
|   V2ADD_RRR_0_OPCODE_X0 = 116,
 | |
|   V2ADD_RRR_0_OPCODE_X1 = 74,
 | |
|   V2ADIFFS_RRR_0_OPCODE_X0 = 117,
 | |
|   V2AVGS_RRR_0_OPCODE_X0 = 118,
 | |
|   V2CMPEQI_IMM8_OPCODE_X0 = 15,
 | |
|   V2CMPEQI_IMM8_OPCODE_X1 = 40,
 | |
|   V2CMPEQ_RRR_0_OPCODE_X0 = 119,
 | |
|   V2CMPEQ_RRR_0_OPCODE_X1 = 75,
 | |
|   V2CMPLES_RRR_0_OPCODE_X0 = 120,
 | |
|   V2CMPLES_RRR_0_OPCODE_X1 = 76,
 | |
|   V2CMPLEU_RRR_0_OPCODE_X0 = 121,
 | |
|   V2CMPLEU_RRR_0_OPCODE_X1 = 77,
 | |
|   V2CMPLTSI_IMM8_OPCODE_X0 = 16,
 | |
|   V2CMPLTSI_IMM8_OPCODE_X1 = 41,
 | |
|   V2CMPLTS_RRR_0_OPCODE_X0 = 122,
 | |
|   V2CMPLTS_RRR_0_OPCODE_X1 = 78,
 | |
|   V2CMPLTUI_IMM8_OPCODE_X0 = 17,
 | |
|   V2CMPLTUI_IMM8_OPCODE_X1 = 42,
 | |
|   V2CMPLTU_RRR_0_OPCODE_X0 = 123,
 | |
|   V2CMPLTU_RRR_0_OPCODE_X1 = 79,
 | |
|   V2CMPNE_RRR_0_OPCODE_X0 = 124,
 | |
|   V2CMPNE_RRR_0_OPCODE_X1 = 80,
 | |
|   V2DOTPA_RRR_0_OPCODE_X0 = 125,
 | |
|   V2DOTP_RRR_0_OPCODE_X0 = 126,
 | |
|   V2INT_H_RRR_0_OPCODE_X0 = 127,
 | |
|   V2INT_H_RRR_0_OPCODE_X1 = 81,
 | |
|   V2INT_L_RRR_0_OPCODE_X0 = 128,
 | |
|   V2INT_L_RRR_0_OPCODE_X1 = 82,
 | |
|   V2MAXSI_IMM8_OPCODE_X0 = 18,
 | |
|   V2MAXSI_IMM8_OPCODE_X1 = 43,
 | |
|   V2MAXS_RRR_0_OPCODE_X0 = 129,
 | |
|   V2MAXS_RRR_0_OPCODE_X1 = 83,
 | |
|   V2MINSI_IMM8_OPCODE_X0 = 19,
 | |
|   V2MINSI_IMM8_OPCODE_X1 = 44,
 | |
|   V2MINS_RRR_0_OPCODE_X0 = 130,
 | |
|   V2MINS_RRR_0_OPCODE_X1 = 84,
 | |
|   V2MNZ_RRR_0_OPCODE_X0 = 131,
 | |
|   V2MNZ_RRR_0_OPCODE_X1 = 85,
 | |
|   V2MULFSC_RRR_0_OPCODE_X0 = 132,
 | |
|   V2MULS_RRR_0_OPCODE_X0 = 133,
 | |
|   V2MULTS_RRR_0_OPCODE_X0 = 134,
 | |
|   V2MZ_RRR_0_OPCODE_X0 = 135,
 | |
|   V2MZ_RRR_0_OPCODE_X1 = 86,
 | |
|   V2PACKH_RRR_0_OPCODE_X0 = 136,
 | |
|   V2PACKH_RRR_0_OPCODE_X1 = 87,
 | |
|   V2PACKL_RRR_0_OPCODE_X0 = 137,
 | |
|   V2PACKL_RRR_0_OPCODE_X1 = 88,
 | |
|   V2PACKUC_RRR_0_OPCODE_X0 = 138,
 | |
|   V2PACKUC_RRR_0_OPCODE_X1 = 89,
 | |
|   V2SADAS_RRR_0_OPCODE_X0 = 139,
 | |
|   V2SADAU_RRR_0_OPCODE_X0 = 140,
 | |
|   V2SADS_RRR_0_OPCODE_X0 = 141,
 | |
|   V2SADU_RRR_0_OPCODE_X0 = 142,
 | |
|   V2SHLI_SHIFT_OPCODE_X0 = 10,
 | |
|   V2SHLI_SHIFT_OPCODE_X1 = 10,
 | |
|   V2SHLSC_RRR_0_OPCODE_X0 = 143,
 | |
|   V2SHLSC_RRR_0_OPCODE_X1 = 90,
 | |
|   V2SHL_RRR_0_OPCODE_X0 = 144,
 | |
|   V2SHL_RRR_0_OPCODE_X1 = 91,
 | |
|   V2SHRSI_SHIFT_OPCODE_X0 = 11,
 | |
|   V2SHRSI_SHIFT_OPCODE_X1 = 11,
 | |
|   V2SHRS_RRR_0_OPCODE_X0 = 145,
 | |
|   V2SHRS_RRR_0_OPCODE_X1 = 92,
 | |
|   V2SHRUI_SHIFT_OPCODE_X0 = 12,
 | |
|   V2SHRUI_SHIFT_OPCODE_X1 = 12,
 | |
|   V2SHRU_RRR_0_OPCODE_X0 = 146,
 | |
|   V2SHRU_RRR_0_OPCODE_X1 = 93,
 | |
|   V2SUBSC_RRR_0_OPCODE_X0 = 147,
 | |
|   V2SUBSC_RRR_0_OPCODE_X1 = 94,
 | |
|   V2SUB_RRR_0_OPCODE_X0 = 148,
 | |
|   V2SUB_RRR_0_OPCODE_X1 = 95,
 | |
|   V4ADDSC_RRR_0_OPCODE_X0 = 149,
 | |
|   V4ADDSC_RRR_0_OPCODE_X1 = 96,
 | |
|   V4ADD_RRR_0_OPCODE_X0 = 150,
 | |
|   V4ADD_RRR_0_OPCODE_X1 = 97,
 | |
|   V4INT_H_RRR_0_OPCODE_X0 = 151,
 | |
|   V4INT_H_RRR_0_OPCODE_X1 = 98,
 | |
|   V4INT_L_RRR_0_OPCODE_X0 = 152,
 | |
|   V4INT_L_RRR_0_OPCODE_X1 = 99,
 | |
|   V4PACKSC_RRR_0_OPCODE_X0 = 153,
 | |
|   V4PACKSC_RRR_0_OPCODE_X1 = 100,
 | |
|   V4SHLSC_RRR_0_OPCODE_X0 = 154,
 | |
|   V4SHLSC_RRR_0_OPCODE_X1 = 101,
 | |
|   V4SHL_RRR_0_OPCODE_X0 = 155,
 | |
|   V4SHL_RRR_0_OPCODE_X1 = 102,
 | |
|   V4SHRS_RRR_0_OPCODE_X0 = 156,
 | |
|   V4SHRS_RRR_0_OPCODE_X1 = 103,
 | |
|   V4SHRU_RRR_0_OPCODE_X0 = 157,
 | |
|   V4SHRU_RRR_0_OPCODE_X1 = 104,
 | |
|   V4SUBSC_RRR_0_OPCODE_X0 = 158,
 | |
|   V4SUBSC_RRR_0_OPCODE_X1 = 105,
 | |
|   V4SUB_RRR_0_OPCODE_X0 = 159,
 | |
|   V4SUB_RRR_0_OPCODE_X1 = 106,
 | |
|   WH64_UNARY_OPCODE_X1 = 38,
 | |
|   XORI_IMM8_OPCODE_X0 = 20,
 | |
|   XORI_IMM8_OPCODE_X1 = 45,
 | |
|   XOR_RRR_0_OPCODE_X0 = 160,
 | |
|   XOR_RRR_0_OPCODE_X1 = 107,
 | |
|   XOR_RRR_5_OPCODE_Y0 = 3,
 | |
|   XOR_RRR_5_OPCODE_Y1 = 3
 | |
| };
 | |
| 
 | |
| 
 | |
| #endif /* __ASSEMBLER__ */
 | |
| 
 | |
| #endif /* OPCODE_TILEGX_H */
 |