res = T0; \
} \
\
-void OPPROTO op_ ## sbc ## l_T0_T1(void) \
-{ \
- res = T0 - T1 + env->CF - 1; \
-} \
- \
void OPPROTO op_ ## sbc ## l_T0_T1_cc(void) \
{ \
unsigned int src1; \
FT0s = vfp_itos(PARAM1);
}
-/* Copy the most significant bit of T0 to all bits of T1. */
-void OPPROTO op_signbit_T1_T0(void)
-{
- T1 = (int32_t)T0 >> 31;
-}
-
void OPPROTO op_movl_cp_T0(void)
{
helper_set_cp(env, PARAM1, T0);
#include "op_addsub.h"
-void OPPROTO op_pkhtb_T0_T1(void)
-{
- T0 = (T0 & 0xffff0000) | (T1 & 0xffff);
-}
-
-void OPPROTO op_pkhbt_T0_T1(void)
-{
- T0 = (T0 & 0xffff) | (T1 & 0xffff0000);
-}
-
-void OPPROTO op_rev16_T0(void)
-{
- T0 = ((T0 & 0xff000000) >> 8)
- | ((T0 & 0x00ff0000) << 8)
- | ((T0 & 0x0000ff00) >> 8)
- | ((T0 & 0x000000ff) << 8);
-}
-
-void OPPROTO op_revsh_T0(void)
-{
- T0 = (int16_t)( ((T0 & 0x0000ff00) >> 8)
- | ((T0 & 0x000000ff) << 8));
-}
-
-void OPPROTO op_rbit_T0(void)
-{
- T0 = ((T0 & 0xff000000) >> 24)
- | ((T0 & 0x00ff0000) >> 8)
- | ((T0 & 0x0000ff00) << 8)
- | ((T0 & 0x000000ff) << 24);
- T0 = ((T0 & 0xf0f0f0f0) >> 4)
- | ((T0 & 0x0f0f0f0f) << 4);
- T0 = ((T0 & 0x88888888) >> 3)
- | ((T0 & 0x44444444) >> 1)
- | ((T0 & 0x22222222) << 1)
- | ((T0 & 0x11111111) << 3);
-}
-
-/* Dual 16-bit signed multiply. */
-void OPPROTO op_mul_dual_T0_T1(void)
-{
- int32_t low;
- int32_t high;
- low = (int32_t)(int16_t)T0 * (int32_t)(int16_t)T1;
- high = (((int32_t)T0) >> 16) * (((int32_t)T1) >> 16);
- T0 = low;
- T1 = high;
-}
-
void OPPROTO op_sel_T0_T1(void)
{
uint32_t mask;
FORCE_RET();
}
-void OPPROTO op_roundqd_T0_T1(void)
-{
- T0 = T1 + ((uint32_t)T0 >> 31);
-}
-
/* Signed saturation. */
static inline uint32_t do_ssat(int32_t val, int shift)
{
T0 = sum;
}
-/* Thumb-2 instructions. */
-
-/* Insert T1 into T0. Result goes in T1. */
-void OPPROTO op_bfi_T1_T0(void)
-{
- int shift = PARAM1;
- uint32_t mask = PARAM2;
- uint32_t bits;
-
- bits = (T1 << shift) & mask;
- T1 = (T0 & ~mask) | bits;
-}
-
-/* Unsigned bitfield extract. */
-void OPPROTO op_ubfx_T1(void)
-{
- uint32_t shift = PARAM1;
- uint32_t mask = PARAM2;
-
- T1 >>= shift;
- T1 &= mask;
-}
-
-/* Signed bitfield extract. */
-void OPPROTO op_sbfx_T1(void)
-{
- uint32_t shift = PARAM1;
- uint32_t width = PARAM2;
- int32_t val;
-
- val = T1 << (32 - (shift + width));
- T1 = val >> (32 - width);
-}
-
-void OPPROTO op_sdivl_T0_T1(void)
-{
- int32_t num;
- int32_t den;
- num = T0;
- den = T1;
- if (den == 0)
- T0 = 0;
- else
- T0 = num / den;
- FORCE_RET();
-}
-
-void OPPROTO op_udivl_T0_T1(void)
-{
- uint32_t num;
- uint32_t den;
- num = T0;
- den = T1;
- if (den == 0)
- T0 = 0;
- else
- T0 = num / den;
- FORCE_RET();
-}
-
void OPPROTO op_movl_T1_r13_banked(void)
{
T1 = helper_get_r13_banked(env, PARAM1);
#define gen_op_subl_T0_T1_usaturate() \
gen_helper_sub_usaturate(cpu_T[0], cpu_T[0], cpu_T[1])
+/* Copy the most significant bit of T0 to all bits of T1. */
+#define gen_op_signbit_T1_T0() tcg_gen_sari_i32(cpu_T[1], cpu_T[0], 31)
+
+static void gen_smul_dual(TCGv a, TCGv b)
+{
+ TCGv tmp1 = new_tmp();
+ TCGv tmp2 = new_tmp();
+ TCGv res;
+ tcg_gen_ext8s_i32(tmp1, a);
+ tcg_gen_ext8s_i32(tmp2, b);
+ tcg_gen_mul_i32(tmp1, tmp1, tmp2);
+ dead_tmp(tmp2);
+ tcg_gen_sari_i32(a, a, 16);
+ tcg_gen_sari_i32(b, b, 16);
+ tcg_gen_mul_i32(b, b, a);
+ tcg_gen_mov_i32(a, tmp1);
+ dead_tmp(tmp1);
+}
+
+/* Byteswap each halfword. */
+static void gen_rev16(TCGv var)
+{
+ TCGv tmp = new_tmp();
+ tcg_gen_shri_i32(tmp, var, 8);
+ tcg_gen_andi_i32(tmp, tmp, 0x00ff00ff);
+ tcg_gen_shli_i32(var, var, 8);
+ tcg_gen_andi_i32(var, var, 0xff00ff00);
+ tcg_gen_or_i32(var, var, tmp);
+ dead_tmp(tmp);
+}
+
+/* Byteswap low halfword and sign extend. */
+static void gen_revsh(TCGv var)
+{
+ TCGv tmp = new_tmp();
+ tcg_gen_shri_i32(tmp, var, 8);
+ tcg_gen_andi_i32(tmp, tmp, 0x00ff);
+ tcg_gen_shli_i32(var, var, 8);
+ tcg_gen_ext8s_i32(var, var);
+ tcg_gen_or_i32(var, var, tmp);
+ dead_tmp(tmp);
+}
+
+/* Unsigned bitfield extract. */
+static void gen_ubfx(TCGv var, int shift, uint32_t mask)
+{
+ if (shift)
+ tcg_gen_shri_i32(var, var, shift);
+ tcg_gen_andi_i32(var, var, mask);
+}
+
+/* Signed bitfield extract. */
+static void gen_sbfx(TCGv var, int shift, int width)
+{
+ uint32_t signbit;
+
+ if (shift)
+ tcg_gen_sari_i32(var, var, shift);
+ if (shift + width < 32) {
+ signbit = 1u << (width - 1);
+ tcg_gen_andi_i32(var, var, (1u << width) - 1);
+ tcg_gen_xori_i32(var, var, signbit);
+ tcg_gen_subi_i32(var, var, signbit);
+ }
+}
+
+/* Bitfield insertion. Insert val into base. Clobbers base and val. */
+static void gen_bfi(TCGv dest, TCGv base, TCGv val, int shift, uint32_t mask)
+{
+ tcg_gen_shli_i32(val, val, shift);
+ tcg_gen_andi_i32(val, val, mask);
+ tcg_gen_andi_i32(base, base, ~mask);
+ tcg_gen_or_i32(dest, base, val);
+}
+
+static void gen_op_roundqd_T0_T1(void)
+{
+ tcg_gen_shri_i32(cpu_T[0], cpu_T[0], 31);
+ tcg_gen_add_i32(cpu_T[0], cpu_T[0], cpu_T[1]);
+}
+
/* FIXME: Most targets have native widening multiplication.
It would be good to use that instead of a full wide multiply. */
/* Unsigned 32x32->64 multiply. */
tcg_gen_shri_i32(tmp, var, 16);
tcg_gen_shli_i32(var, var, 16);
tcg_gen_or_i32(var, var, tmp);
+ dead_tmp(tmp);
}
/* Dual 16-bit add. Result placed in t0 and t1 is marked as dead.
dead_tmp(tmp);
}
+/* dest = T0 - T1 + CF - 1. */
+static void gen_sub_carry(TCGv dest, TCGv t0, TCGv t1)
+{
+ TCGv tmp = new_tmp();
+ tcg_gen_sub_i32(dest, t0, t1);
+ tcg_gen_ld_i32(tmp, cpu_env, offsetof(CPUState, CF));
+ tcg_gen_add_i32(dest, dest, tmp);
+ tcg_gen_subi_i32(dest, dest, 1);
+ dead_tmp(tmp);
+}
+
+#define gen_sbc_T0_T1() gen_sub_carry(cpu_T[0], cpu_T[0], cpu_T[1])
+#define gen_rsc_T0_T1() gen_sub_carry(cpu_T[0], cpu_T[1], cpu_T[0])
+
/* FIXME: Implement this natively. */
static inline void tcg_gen_not_i32(TCGv t0, TCGv t1)
{
case 2: /* VREV16 */
if (size != 0)
return 1;
- gen_op_rev16_T0();
+ gen_rev16(cpu_T[0]);
break;
case 4: case 5: /* VPADDL */
case 12: case 13: /* VPADAL */
{
unsigned int cond, insn, val, op1, i, shift, rm, rs, rn, rd, sh;
TCGv tmp;
+ TCGv tmp2;
insn = ldl_code(s->pc);
s->pc += 4;
if (set_cc)
gen_op_sbcl_T0_T1_cc();
else
- gen_op_sbcl_T0_T1();
+ gen_sbc_T0_T1();
gen_movl_reg_T0(s, rd);
break;
case 0x07:
if (set_cc)
gen_op_rscl_T0_T1_cc();
else
- gen_op_rscl_T0_T1();
+ gen_rsc_T0_T1();
gen_movl_reg_T0(s, rd);
break;
case 0x08:
case 1:
if ((insn & 0x00700020) == 0) {
/* Hafword pack. */
- gen_movl_T0_reg(s, rn);
- gen_movl_T1_reg(s, rm);
+ tmp = load_reg(s, rn);
+ tmp2 = load_reg(s, rm);
shift = (insn >> 7) & 0x1f;
if (shift)
- gen_op_shll_T1_im(shift);
- if (insn & (1 << 6))
- gen_op_pkhtb_T0_T1();
- else
- gen_op_pkhbt_T0_T1();
- gen_movl_reg_T0(s, rd);
+ tcg_gen_shli_i32(tmp2, tmp2, shift);
+ if (insn & (1 << 6)) {
+ /* pkhtb */
+ tcg_gen_andi_i32(tmp, tmp, 0xffff0000);
+ tcg_gen_andi_i32(tmp2, tmp2, 0xffff);
+ } else {
+ /* pkhbt */
+ tcg_gen_andi_i32(tmp, tmp, 0xffff);
+ tcg_gen_andi_i32(tmp2, tmp2, 0xffff0000);
+ }
+ tcg_gen_or_i32(tmp, tmp, tmp2);
+ store_reg(s, rd, tmp);
} else if ((insn & 0x00200020) == 0x00200000) {
/* [us]sat */
gen_movl_T1_reg(s, rm);
gen_movl_T0_reg(s, rm);
if (insn & (1 << 22)) {
if (insn & (1 << 7)) {
- gen_op_revsh_T0();
+ gen_revsh(cpu_T[0]);
} else {
ARCH(6T2);
- gen_op_rbit_T0();
+ gen_helper_rbit(cpu_T[0], cpu_T[0]);
}
} else {
if (insn & (1 << 7))
- gen_op_rev16_T0();
+ gen_rev16(cpu_T[0]);
else
gen_op_rev_T0();
}
} else {
if (insn & (1 << 5))
gen_swap_half(cpu_T[1]);
- gen_op_mul_dual_T0_T1();
+ gen_smul_dual(cpu_T[0], cpu_T[1]);
if (insn & (1 << 22)) {
if (insn & (1 << 6)) {
/* smlald */
}
if (i != 32) {
gen_movl_T0_reg(s, rd);
- gen_op_bfi_T1_T0(shift, ((1u << i) - 1) << shift);
+ gen_bfi(cpu_T[1], cpu_T[0], cpu_T[1],
+ shift, ((1u << i) - 1) << shift);
}
gen_movl_reg_T1(s, rd);
break;
goto illegal_op;
if (i < 32) {
if (op1 & 0x20) {
- gen_op_ubfx_T1(shift, (1u << i) - 1);
+ gen_ubfx(cpu_T[1], shift, (1u << i) - 1);
} else {
- gen_op_sbfx_T1(shift, i);
+ gen_sbfx(cpu_T[1], shift, i);
}
}
gen_movl_reg_T1(s, rd);
if (conds)
gen_op_sbcl_T0_T1_cc();
else
- gen_op_sbcl_T0_T1();
+ gen_sbc_T0_T1();
break;
case 13: /* sub */
if (conds)
gen_movl_T0_reg(s, rn);
switch (op) {
case 0x0a: /* rbit */
- gen_op_rbit_T0();
+ gen_helper_rbit(cpu_T[0], cpu_T[0]);
break;
case 0x08: /* rev */
gen_op_rev_T0();
break;
case 0x09: /* rev16 */
- gen_op_rev16_T0();
+ gen_rev16(cpu_T[0]);
break;
case 0x0b: /* revsh */
- gen_op_revsh_T0();
+ gen_revsh(cpu_T[0]);
break;
case 0x10: /* sel */
gen_movl_T1_reg(s, rm);
case 4: /* Dual multiply subtract. */
if (op)
gen_swap_half(cpu_T[1]);
- gen_op_mul_dual_T0_T1();
+ gen_smul_dual(cpu_T[0], cpu_T[1]);
/* This addition cannot overflow. */
if (insn & (1 << 22)) {
gen_op_subl_T0_T1();
if (!arm_feature(env, ARM_FEATURE_DIV))
goto illegal_op;
if (op & 0x20)
- gen_op_udivl_T0_T1();
+ gen_helper_udiv(cpu_T[0], cpu_T[0], cpu_T[1]);
else
- gen_op_sdivl_T0_T1();
+ gen_helper_sdiv(cpu_T[0], cpu_T[0], cpu_T[1]);
gen_movl_reg_T0(s, rd);
} else if ((op & 0xe) == 0xc) {
/* Dual multiply accumulate long. */
if (op & 1)
gen_swap_half(cpu_T[1]);
- gen_op_mul_dual_T0_T1();
+ gen_smul_dual(cpu_T[0], cpu_T[1]);
if (op & 0x10) {
gen_op_subl_T0_T1();
} else {
if (shift + imm > 32)
goto illegal_op;
if (imm < 32)
- gen_op_sbfx_T1(shift, imm);
+ gen_sbfx(cpu_T[1], shift, imm);
break;
case 6: /* Unsigned bitfield extract. */
imm++;
if (shift + imm > 32)
goto illegal_op;
if (imm < 32)
- gen_op_ubfx_T1(shift, (1u << imm) - 1);
+ gen_ubfx(cpu_T[1], shift, (1u << imm) - 1);
break;
case 3: /* Bitfield insert/clear. */
if (imm < shift)
imm = imm + 1 - shift;
if (imm != 32) {
gen_movl_T0_reg(s, rd);
- gen_op_bfi_T1_T0(shift, ((1u << imm) - 1) << shift);
+ gen_bfi(cpu_T[1], cpu_T[0], cpu_T[1],
+ shift, ((1u << imm) - 1) << shift);
}
break;
case 7:
break;
case 0x6: /* sbc */
if (s->condexec_mask)
- gen_op_sbcl_T0_T1();
+ gen_sbc_T0_T1();
else
gen_op_sbcl_T0_T1_cc();
break;
gen_movl_T0_reg(s, rn);
switch ((insn >> 6) & 3) {
case 0: gen_op_rev_T0(); break;
- case 1: gen_op_rev16_T0(); break;
- case 3: gen_op_revsh_T0(); break;
+ case 1: gen_rev16(cpu_T[0]); break;
+ case 3: gen_revsh(cpu_T[0]); break;
default: goto illegal_op;
}
gen_movl_reg_T0(s, rd);