|
|
|
@@ -126,6 +126,7 @@ static bool patch_reloc(tcg_insn_unit *code_ptr, int type,
|
|
|
|
|
#define TCG_CT_CONST_MONE 0x800
|
|
|
|
|
#define TCG_CT_CONST_ORRI 0x1000
|
|
|
|
|
#define TCG_CT_CONST_ANDI 0x2000
|
|
|
|
|
#define TCG_CT_CONST_CMP 0x4000
|
|
|
|
|
|
|
|
|
|
#define ALL_GENERAL_REGS 0xffffffffu
|
|
|
|
|
#define ALL_VECTOR_REGS 0xffffffff00000000ull
|
|
|
|
@@ -279,6 +280,15 @@ static bool tcg_target_const_match(int64_t val, int ct,
|
|
|
|
|
if (type == TCG_TYPE_I32) {
|
|
|
|
|
val = (int32_t)val;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (ct & TCG_CT_CONST_CMP) {
|
|
|
|
|
if (is_tst_cond(cond)) {
|
|
|
|
|
ct |= TCG_CT_CONST_LIMM;
|
|
|
|
|
} else {
|
|
|
|
|
ct |= TCG_CT_CONST_AIMM;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if ((ct & TCG_CT_CONST_AIMM) && (is_aimm(val) || is_aimm(-val))) {
|
|
|
|
|
return 1;
|
|
|
|
|
}
|
|
|
|
@@ -345,6 +355,9 @@ static const enum aarch64_cond_code tcg_cond_to_aarch64[] = {
|
|
|
|
|
[TCG_COND_GTU] = COND_HI,
|
|
|
|
|
[TCG_COND_GEU] = COND_HS,
|
|
|
|
|
[TCG_COND_LEU] = COND_LS,
|
|
|
|
|
/* bit test */
|
|
|
|
|
[TCG_COND_TSTEQ] = COND_EQ,
|
|
|
|
|
[TCG_COND_TSTNE] = COND_NE,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
typedef enum {
|
|
|
|
@@ -1342,19 +1355,26 @@ static inline void tcg_out_dep(TCGContext *s, TCGType ext, TCGReg rd,
|
|
|
|
|
tcg_out_bfm(s, ext, rd, rn, a, b);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void tcg_out_cmp(TCGContext *s, TCGType ext, TCGReg a,
|
|
|
|
|
static void tcg_out_cmp(TCGContext *s, TCGType ext, TCGCond cond, TCGReg a,
|
|
|
|
|
tcg_target_long b, bool const_b)
|
|
|
|
|
{
|
|
|
|
|
if (const_b) {
|
|
|
|
|
/* Using CMP or CMN aliases. */
|
|
|
|
|
if (b >= 0) {
|
|
|
|
|
tcg_out_insn(s, 3401, SUBSI, ext, TCG_REG_XZR, a, b);
|
|
|
|
|
if (is_tst_cond(cond)) {
|
|
|
|
|
if (!const_b) {
|
|
|
|
|
tcg_out_insn(s, 3510, ANDS, ext, TCG_REG_XZR, a, b);
|
|
|
|
|
} else {
|
|
|
|
|
tcg_out_insn(s, 3401, ADDSI, ext, TCG_REG_XZR, a, -b);
|
|
|
|
|
tcg_debug_assert(is_limm(b));
|
|
|
|
|
tcg_out_logicali(s, I3404_ANDSI, 0, TCG_REG_XZR, a, b);
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
/* Using CMP alias SUBS wzr, Wn, Wm */
|
|
|
|
|
tcg_out_insn(s, 3502, SUBS, ext, TCG_REG_XZR, a, b);
|
|
|
|
|
if (!const_b) {
|
|
|
|
|
tcg_out_insn(s, 3502, SUBS, ext, TCG_REG_XZR, a, b);
|
|
|
|
|
} else if (b >= 0) {
|
|
|
|
|
tcg_debug_assert(is_aimm(b));
|
|
|
|
|
tcg_out_insn(s, 3401, SUBSI, ext, TCG_REG_XZR, a, b);
|
|
|
|
|
} else {
|
|
|
|
|
tcg_debug_assert(is_aimm(-b));
|
|
|
|
|
tcg_out_insn(s, 3401, ADDSI, ext, TCG_REG_XZR, a, -b);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
@@ -1402,7 +1422,7 @@ static void tcg_out_brcond(TCGContext *s, TCGType ext, TCGCond c, TCGArg a,
|
|
|
|
|
need_cmp = false;
|
|
|
|
|
} else {
|
|
|
|
|
need_cmp = true;
|
|
|
|
|
tcg_out_cmp(s, ext, a, b, b_const);
|
|
|
|
|
tcg_out_cmp(s, ext, c, a, b, b_const);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!l->has_value) {
|
|
|
|
@@ -1575,7 +1595,7 @@ static void tcg_out_cltz(TCGContext *s, TCGType ext, TCGReg d,
|
|
|
|
|
} else {
|
|
|
|
|
AArch64Insn sel = I3506_CSEL;
|
|
|
|
|
|
|
|
|
|
tcg_out_cmp(s, ext, a0, 0, 1);
|
|
|
|
|
tcg_out_cmp(s, ext, TCG_COND_NE, a0, 0, 1);
|
|
|
|
|
tcg_out_insn(s, 3507, CLZ, ext, TCG_REG_TMP0, a1);
|
|
|
|
|
|
|
|
|
|
if (const_b) {
|
|
|
|
@@ -1720,7 +1740,7 @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h,
|
|
|
|
|
addr_adj, compare_mask);
|
|
|
|
|
|
|
|
|
|
/* Perform the address comparison. */
|
|
|
|
|
tcg_out_cmp(s, addr_type, TCG_REG_TMP0, TCG_REG_TMP2, 0);
|
|
|
|
|
tcg_out_cmp(s, addr_type, TCG_COND_NE, TCG_REG_TMP0, TCG_REG_TMP2, 0);
|
|
|
|
|
|
|
|
|
|
/* If not equal, we jump to the slow path. */
|
|
|
|
|
ldst->label_ptr[0] = s->code_ptr;
|
|
|
|
@@ -2276,7 +2296,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
|
|
|
|
|
a2 = (int32_t)a2;
|
|
|
|
|
/* FALLTHRU */
|
|
|
|
|
case INDEX_op_setcond_i64:
|
|
|
|
|
tcg_out_cmp(s, ext, a1, a2, c2);
|
|
|
|
|
tcg_out_cmp(s, ext, args[3], a1, a2, c2);
|
|
|
|
|
/* Use CSET alias of CSINC Wd, WZR, WZR, invert(cond). */
|
|
|
|
|
tcg_out_insn(s, 3506, CSINC, TCG_TYPE_I32, a0, TCG_REG_XZR,
|
|
|
|
|
TCG_REG_XZR, tcg_invert_cond(args[3]));
|
|
|
|
@@ -2286,7 +2306,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
|
|
|
|
|
a2 = (int32_t)a2;
|
|
|
|
|
/* FALLTHRU */
|
|
|
|
|
case INDEX_op_negsetcond_i64:
|
|
|
|
|
tcg_out_cmp(s, ext, a1, a2, c2);
|
|
|
|
|
tcg_out_cmp(s, ext, args[3], a1, a2, c2);
|
|
|
|
|
/* Use CSETM alias of CSINV Wd, WZR, WZR, invert(cond). */
|
|
|
|
|
tcg_out_insn(s, 3506, CSINV, ext, a0, TCG_REG_XZR,
|
|
|
|
|
TCG_REG_XZR, tcg_invert_cond(args[3]));
|
|
|
|
@@ -2296,7 +2316,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
|
|
|
|
|
a2 = (int32_t)a2;
|
|
|
|
|
/* FALLTHRU */
|
|
|
|
|
case INDEX_op_movcond_i64:
|
|
|
|
|
tcg_out_cmp(s, ext, a1, a2, c2);
|
|
|
|
|
tcg_out_cmp(s, ext, args[5], a1, a2, c2);
|
|
|
|
|
tcg_out_insn(s, 3506, CSEL, ext, a0, REG0(3), REG0(4), args[5]);
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
@@ -2896,11 +2916,13 @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
|
|
|
|
|
case INDEX_op_add_i64:
|
|
|
|
|
case INDEX_op_sub_i32:
|
|
|
|
|
case INDEX_op_sub_i64:
|
|
|
|
|
return C_O1_I2(r, r, rA);
|
|
|
|
|
|
|
|
|
|
case INDEX_op_setcond_i32:
|
|
|
|
|
case INDEX_op_setcond_i64:
|
|
|
|
|
case INDEX_op_negsetcond_i32:
|
|
|
|
|
case INDEX_op_negsetcond_i64:
|
|
|
|
|
return C_O1_I2(r, r, rA);
|
|
|
|
|
return C_O1_I2(r, r, rC);
|
|
|
|
|
|
|
|
|
|
case INDEX_op_mul_i32:
|
|
|
|
|
case INDEX_op_mul_i64:
|
|
|
|
@@ -2950,11 +2972,11 @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
|
|
|
|
|
|
|
|
|
|
case INDEX_op_brcond_i32:
|
|
|
|
|
case INDEX_op_brcond_i64:
|
|
|
|
|
return C_O0_I2(r, rA);
|
|
|
|
|
return C_O0_I2(r, rC);
|
|
|
|
|
|
|
|
|
|
case INDEX_op_movcond_i32:
|
|
|
|
|
case INDEX_op_movcond_i64:
|
|
|
|
|
return C_O1_I4(r, r, rA, rZ, rZ);
|
|
|
|
|
return C_O1_I4(r, r, rC, rZ, rZ);
|
|
|
|
|
|
|
|
|
|
case INDEX_op_qemu_ld_a32_i32:
|
|
|
|
|
case INDEX_op_qemu_ld_a64_i32:
|
|
|
|
|