diff options
author | pbrook <pbrook@c046a42c-6fe2-441c-8c8c-71466251a162> | 2006-10-22 00:18:54 +0000 |
---|---|---|
committer | pbrook <pbrook@c046a42c-6fe2-441c-8c8c-71466251a162> | 2006-10-22 00:18:54 +0000 |
commit | e6e5906b6e0a81718066ca43aef57515026c6624 (patch) | |
tree | 79996f66ed5a2e4f37783a114cb45f6913204486 /target-m68k/translate.c | |
parent | 223b8a40d13550ebc08216bd96d1f615597e0554 (diff) | |
download | qemu-e6e5906b6e0a81718066ca43aef57515026c6624.zip qemu-e6e5906b6e0a81718066ca43aef57515026c6624.tar.gz qemu-e6e5906b6e0a81718066ca43aef57515026c6624.tar.bz2 |
ColdFire target.
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@2196 c046a42c-6fe2-441c-8c8c-71466251a162
Diffstat (limited to 'target-m68k/translate.c')
-rw-r--r-- | target-m68k/translate.c | 2759 |
1 files changed, 2759 insertions, 0 deletions
diff --git a/target-m68k/translate.c b/target-m68k/translate.c new file mode 100644 index 0000000..10728f9 --- /dev/null +++ b/target-m68k/translate.c @@ -0,0 +1,2759 @@ +/* + * m68k translation + * + * Copyright (c) 2005-2006 CodeSourcery + * Written by Paul Brook + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with this library; if not, write to the Free Software + * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA + */ +#include <stdarg.h> +#include <stdlib.h> +#include <stdio.h> +#include <string.h> +#include <inttypes.h> + +#include "config.h" +#include "cpu.h" +#include "exec-all.h" +#include "disas.h" +#include "m68k-qreg.h" + +static inline void qemu_assert(int cond, const char *msg) +{ + if (!cond) { + fprintf (stderr, "badness: %s\n", msg); + abort(); + } +} + +/* internal defines */ +typedef struct DisasContext { + target_ulong pc; + int is_jmp; + int cc_op; + uint32_t fpcr; + struct TranslationBlock *tb; + int singlestep_enabled; +} DisasContext; + +#define DISAS_JUMP_NEXT 4 + +/* XXX: move that elsewhere */ +/* ??? Fix exceptions. */ +static void *gen_throws_exception; +#define gen_last_qop NULL + +static uint16_t *gen_opc_ptr; +static uint32_t *gen_opparam_ptr; +extern FILE *logfile; +extern int loglevel; + +enum { +#define DEF(s, n, copy_size) INDEX_op_ ## s, +#include "opc.h" +#undef DEF + NB_OPS, +}; + +#include "gen-op.h" +#include "op-hacks.h" + +#define OS_BYTE 0 +#define OS_WORD 1 +#define OS_LONG 2 +#define OS_SINGLE 4 +#define OS_DOUBLE 5 + +#define DREG(insn, pos) (((insn >> pos) & 7) + QREG_D0) +#define AREG(insn, pos) (((insn >> pos) & 7) + QREG_A0) +#define FREG(insn, pos) (((insn >> pos) & 7) + QREG_F0) + +#define M68K_INSN_CF_A (1 << 0) +#define M68K_INSN_CF_B (1 << 1) +#define M68K_INSN_CF_C (1 << 2) +#define M68K_INSN_CF_MAC (1 << 3) +#define M68K_INSN_CF_EMAC (1 << 4) +#define M68K_INSN_CF_FPU (1 << 5) + +struct m68k_def_t { + const char * name; + uint32_t insns; +}; + +static m68k_def_t m68k_cpu_defs[] = { + {"m5206", M68K_INSN_CF_A}, + {"cfv4e", M68K_INSN_CF_A | M68K_INSN_CF_B | M68K_INSN_CF_C + | M68K_INSN_CF_MAC | M68K_INSN_CF_EMAC | M68K_INSN_CF_FPU}, + {NULL, 0}, +}; + +typedef void (*disas_proc)(DisasContext *, uint16_t); + +#define DISAS_INSN(name) \ + static void disas_##name (DisasContext *s, uint16_t insn) + +/* Generate a load from the specified address. Narrow values are + sign extended to full register width. */ +static inline int gen_load(int opsize, int addr, int sign) +{ + int tmp; + switch(opsize) { + case OS_BYTE: + tmp = gen_new_qreg(QMODE_I32); + if (sign) + gen_op_ld8s32(tmp, addr); + else + gen_op_ld8u32(tmp, addr); + break; + case OS_WORD: + tmp = gen_new_qreg(QMODE_I32); + if (sign) + gen_op_ld16s32(tmp, addr); + else + gen_op_ld16u32(tmp, addr); + break; + case OS_LONG: + tmp = gen_new_qreg(QMODE_I32); + gen_op_ld32(tmp, addr); + break; + case OS_SINGLE: + tmp = gen_new_qreg(QMODE_F32); + gen_op_ldf32(tmp, addr); + break; + case OS_DOUBLE: + tmp = gen_new_qreg(QMODE_F64); + gen_op_ldf64(tmp, addr); + break; + default: + qemu_assert(0, "bad load size"); + } + gen_throws_exception = gen_last_qop; + return tmp; +} + +/* Generate a store. */ +static inline void gen_store(int opsize, int addr, int val) +{ + switch(opsize) { + case OS_BYTE: + gen_op_st8(addr, val); + break; + case OS_WORD: + gen_op_st16(addr, val); + break; + case OS_LONG: + gen_op_st32(addr, val); + break; + case OS_SINGLE: + gen_op_stf32(addr, val); + break; + case OS_DOUBLE: + gen_op_stf64(addr, val); + break; + default: + qemu_assert(0, "bad store size"); + } + gen_throws_exception = gen_last_qop; +} + +/* Generate an unsigned load if VAL is 0 a signed load if val is -1, + otherwise generate a store. */ +static int gen_ldst(int opsize, int addr, int val) +{ + if (val > 0) { + gen_store(opsize, addr, val); + return 0; + } else { + return gen_load(opsize, addr, val != 0); + } +} + +/* Handle a base + index + displacement effective addresss. A base of + -1 means pc-relative. */ +static int gen_lea_indexed(DisasContext *s, int opsize, int base) +{ + int scale; + uint32_t offset; + uint16_t ext; + int add; + int tmp; + + offset = s->pc; + ext = lduw(s->pc); + s->pc += 2; + tmp = ((ext >> 12) & 7) + ((ext & 0x8000) ? QREG_A0 : QREG_D0); + /* ??? Check W/L bit. */ + scale = (ext >> 9) & 3; + if (scale == 0) { + add = tmp; + } else { + add = gen_new_qreg(QMODE_I32); + gen_op_shl32(add, tmp, gen_im32(scale)); + } + tmp = gen_new_qreg(QMODE_I32); + if (base != -1) { + gen_op_add32(tmp, base, gen_im32((int8_t)ext)); + gen_op_add32(tmp, tmp, add); + } else { + gen_op_add32(tmp, add, gen_im32(offset + (int8_t)ext)); + } + return tmp; +} + +/* Read a 32-bit immediate constant. */ +static inline uint32_t read_im32(DisasContext *s) +{ + uint32_t im; + im = ((uint32_t)lduw(s->pc)) << 16; + s->pc += 2; + im |= lduw(s->pc); + s->pc += 2; + return im; +} + + +/* Update the CPU env CC_OP state. */ +static inline void gen_flush_cc_op(DisasContext *s) +{ + if (s->cc_op != CC_OP_DYNAMIC) + gen_op_mov32(QREG_CC_OP, gen_im32(s->cc_op)); +} + +/* Evaluate all the CC flags. */ +static inline void gen_flush_flags(DisasContext *s) +{ + if (s->cc_op == CC_OP_FLAGS) + return; + gen_op_flush_flags(s->cc_op); + s->cc_op = CC_OP_FLAGS; +} + +static inline int opsize_bytes(int opsize) +{ + switch (opsize) { + case OS_BYTE: return 1; + case OS_WORD: return 2; + case OS_LONG: return 4; + case OS_SINGLE: return 4; + case OS_DOUBLE: return 8; + default: + qemu_assert(0, "bad operand size"); + } +} + +/* Assign value to a register. If the width is less than the register width + only the low part of the register is set. */ +static void gen_partset_reg(int opsize, int reg, int val) +{ + int tmp; + switch (opsize) { + case OS_BYTE: + gen_op_and32(reg, reg, gen_im32(0xffffff00)); + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, val, gen_im32(0xff)); + gen_op_or32(reg, reg, tmp); + break; + case OS_WORD: + gen_op_and32(reg, reg, gen_im32(0xffff0000)); + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, val, gen_im32(0xffff)); + gen_op_or32(reg, reg, tmp); + break; + case OS_LONG: + gen_op_mov32(reg, val); + break; + case OS_SINGLE: + gen_op_pack_32_f32(reg, val); + break; + default: + qemu_assert(0, "Bad operand size"); + break; + } +} + +/* Sign or zero extend a value. */ +static inline int gen_extend(int val, int opsize, int sign) +{ + int tmp; + + switch (opsize) { + case OS_BYTE: + tmp = gen_new_qreg(QMODE_I32); + if (sign) + gen_op_ext8s32(tmp, val); + else + gen_op_ext8u32(tmp, val); + break; + case OS_WORD: + tmp = gen_new_qreg(QMODE_I32); + if (sign) + gen_op_ext16s32(tmp, val); + else + gen_op_ext16u32(tmp, val); + break; + case OS_LONG: + tmp = val; + break; + case OS_SINGLE: + tmp = gen_new_qreg(QMODE_F32); + gen_op_pack_f32_32(tmp, val); + break; + default: + qemu_assert(0, "Bad operand size"); + } + return tmp; +} + +/* Generate code for an "effective address". Does not adjust the base + register for autoincrememnt addressing modes. */ +static int gen_lea(DisasContext *s, uint16_t insn, int opsize) +{ + int reg; + int tmp; + uint16_t ext; + uint32_t offset; + + reg = insn & 7; + switch ((insn >> 3) & 7) { + case 0: /* Data register direct. */ + case 1: /* Address register direct. */ + /* ??? generate bad addressing mode fault. */ + qemu_assert(0, "invalid addressing mode"); + case 2: /* Indirect register */ + case 3: /* Indirect postincrement. */ + reg += QREG_A0; + return reg; + case 4: /* Indirect predecrememnt. */ + reg += QREG_A0; + tmp = gen_new_qreg(QMODE_I32); + gen_op_sub32(tmp, reg, gen_im32(opsize_bytes(opsize))); + return tmp; + case 5: /* Indirect displacement. */ + reg += QREG_A0; + tmp = gen_new_qreg(QMODE_I32); + ext = lduw(s->pc); + s->pc += 2; + gen_op_add32(tmp, reg, gen_im32((int16_t)ext)); + return tmp; + case 6: /* Indirect index + displacement. */ + reg += QREG_A0; + return gen_lea_indexed(s, opsize, reg); + case 7: /* Other */ + switch (reg) { + case 0: /* Absolute short. */ + offset = ldsw(s->pc); + s->pc += 2; + return gen_im32(offset); + case 1: /* Absolute long. */ + offset = read_im32(s); + return gen_im32(offset); + case 2: /* pc displacement */ + tmp = gen_new_qreg(QMODE_I32); + offset = s->pc; + offset += ldsw(s->pc); + s->pc += 2; + return gen_im32(offset); + case 3: /* pc index+displacement. */ + return gen_lea_indexed(s, opsize, -1); + case 4: /* Immediate. */ + default: + /* ??? generate bad addressing mode fault. */ + qemu_assert(0, "invalid addressing mode"); + } + } + /* Should never happen. */ + return -1; +} + +/* Helper function for gen_ea. Reuse the computed address between the + for read/write operands. */ +static inline int gen_ea_once(DisasContext *s, uint16_t insn, int opsize, + int val, int *addrp) +{ + int tmp; + + if (addrp && val > 0) { + tmp = *addrp; + } else { + tmp = gen_lea(s, insn, opsize); + if (addrp) + *addrp = tmp; + } + return gen_ldst(opsize, tmp, val); +} + +/* Generate code to load/store a value ito/from an EA. If VAL > 0 this is + a write otherwise it is a read (0 == sign extend, -1 == zero extend). + ADDRP is non-null for readwrite operands. */ +static int gen_ea(DisasContext *s, uint16_t insn, int opsize, int val, + int *addrp) +{ + int reg; + int result; + uint32_t offset; + + reg = insn & 7; + switch ((insn >> 3) & 7) { + case 0: /* Data register direct. */ + reg += QREG_D0; + if (val > 0) { + gen_partset_reg(opsize, reg, val); + return 0; + } else { + return gen_extend(reg, opsize, val); + } + case 1: /* Address register direct. */ + reg += QREG_A0; + if (val > 0) { + gen_op_mov32(reg, val); + return 0; + } else { + return gen_extend(reg, opsize, val); + } + case 2: /* Indirect register */ + reg += QREG_A0; + return gen_ldst(opsize, reg, val); + case 3: /* Indirect postincrement. */ + reg += QREG_A0; + result = gen_ldst(opsize, reg, val); + /* ??? This is not exception safe. The instruction may still + fault after this point. */ + if (val > 0 || !addrp) + gen_op_add32(reg, reg, gen_im32(opsize_bytes(opsize))); + return result; + case 4: /* Indirect predecrememnt. */ + { + int tmp; + if (addrp && val > 0) { + tmp = *addrp; + } else { + tmp = gen_lea(s, insn, opsize); + if (addrp) + *addrp = tmp; + } + result = gen_ldst(opsize, tmp, val); + /* ??? This is not exception safe. The instruction may still + fault after this point. */ + if (val > 0 || !addrp) { + reg += QREG_A0; + gen_op_mov32(reg, tmp); + } + } + return result; + case 5: /* Indirect displacement. */ + case 6: /* Indirect index + displacement. */ + return gen_ea_once(s, insn, opsize, val, addrp); + case 7: /* Other */ + switch (reg) { + case 0: /* Absolute short. */ + case 1: /* Absolute long. */ + case 2: /* pc displacement */ + case 3: /* pc index+displacement. */ + return gen_ea_once(s, insn, opsize, val, addrp); + case 4: /* Immediate. */ + /* Sign extend values for consistency. */ + switch (opsize) { + case OS_BYTE: + if (val) + offset = ldsb(s->pc + 1); + else + offset = ldub(s->pc + 1); + s->pc += 2; + break; + case OS_WORD: + if (val) + offset = ldsw(s->pc); + else + offset = lduw(s->pc); + s->pc += 2; + break; + case OS_LONG: + offset = read_im32(s); + break; + default: + qemu_assert(0, "Bad immediate operand"); + } + return gen_im32(offset); + default: + qemu_assert(0, "invalid addressing mode"); + } + } + /* Should never happen. */ + return -1; +} + +static void gen_logic_cc(DisasContext *s, int val) +{ + gen_op_logic_cc(val); + s->cc_op = CC_OP_LOGIC; +} + +static void gen_jmpcc(DisasContext *s, int cond, int l1) +{ + int tmp; + + gen_flush_flags(s); + switch (cond) { + case 0: /* T */ + gen_op_jmp(l1); + break; + case 1: /* F */ + break; + case 2: /* HI (!C && !Z) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_C | CCF_Z)); + gen_op_jmp_z32(tmp, l1); + break; + case 3: /* LS (C || Z) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_C | CCF_Z)); + gen_op_jmp_nz32(tmp, l1); + break; + case 4: /* CC (!C) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_C)); + gen_op_jmp_z32(tmp, l1); + break; + case 5: /* CS (C) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_C)); + gen_op_jmp_nz32(tmp, l1); + break; + case 6: /* NE (!Z) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_Z)); + gen_op_jmp_z32(tmp, l1); + break; + case 7: /* EQ (Z) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_Z)); + gen_op_jmp_nz32(tmp, l1); + break; + case 8: /* VC (!V) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_V)); + gen_op_jmp_z32(tmp, l1); + break; + case 9: /* VS (V) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_V)); + gen_op_jmp_nz32(tmp, l1); + break; + case 10: /* PL (!N) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_N)); + gen_op_jmp_z32(tmp, l1); + break; + case 11: /* MI (N) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_N)); + gen_op_jmp_nz32(tmp, l1); + break; + case 12: /* GE (!(N ^ V)) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_shr32(tmp, QREG_CC_DEST, gen_im32(2)); + gen_op_xor32(tmp, tmp, QREG_CC_DEST); + gen_op_and32(tmp, tmp, gen_im32(CCF_V)); + gen_op_jmp_z32(tmp, l1); + break; + case 13: /* LT (N ^ V) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_shr32(tmp, QREG_CC_DEST, gen_im32(2)); + gen_op_xor32(tmp, tmp, QREG_CC_DEST); + gen_op_and32(tmp, tmp, gen_im32(CCF_V)); + gen_op_jmp_nz32(tmp, l1); + break; + case 14: /* GT (!(Z || (N ^ V))) */ + { + int l2; + l2 = gen_new_label(); + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_Z)); + gen_op_jmp_nz32(tmp, l2); + tmp = gen_new_qreg(QMODE_I32); + gen_op_shr32(tmp, QREG_CC_DEST, gen_im32(2)); + gen_op_xor32(tmp, tmp, QREG_CC_DEST); + gen_op_and32(tmp, tmp, gen_im32(CCF_V)); + gen_op_jmp_nz32(tmp, l2); + gen_op_jmp(l1); + gen_set_label(l2); + } + break; + case 15: /* LE (Z || (N ^ V)) */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_Z)); + gen_op_jmp_nz32(tmp, l1); + tmp = gen_new_qreg(QMODE_I32); + gen_op_shr32(tmp, QREG_CC_DEST, gen_im32(2)); + gen_op_xor32(tmp, tmp, QREG_CC_DEST); + gen_op_and32(tmp, tmp, gen_im32(CCF_V)); + gen_op_jmp_nz32(tmp, l1); + break; + default: + /* Should ever happen. */ + abort(); + } +} + +DISAS_INSN(scc) +{ + int l1; + int cond; + int reg; + + l1 = gen_new_label(); + cond = (insn >> 8) & 0xf; + reg = DREG(insn, 0); + gen_op_and32(reg, reg, gen_im32(0xffffff00)); + gen_jmpcc(s, cond ^ 1, l1); + gen_op_or32(reg, reg, gen_im32(0xff)); + gen_set_label(l1); +} + +/* Generate a jump to to the address in qreg DEST. */ +static void gen_jmp(DisasContext *s, int dest) +{ + gen_flush_cc_op(s); + gen_op_mov32(QREG_PC, dest); + s->is_jmp = DISAS_JUMP; +} + +static void gen_exception(DisasContext *s, uint32_t where, int nr) +{ + gen_flush_cc_op(s); + gen_jmp(s, gen_im32(where)); + gen_op_raise_exception(nr); +} + +/* Generate a jump to an immediate address. */ +static void gen_jmp_tb(DisasContext *s, int n, uint32_t dest) +{ + TranslationBlock *tb; + + tb = s->tb; + if (__builtin_expect (s->singlestep_enabled, 0)) { + gen_exception(s, dest, EXCP_DEBUG); + } else if ((tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) || + (s->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK)) { + gen_op_goto_tb(0, n, (long)tb); + gen_op_mov32(QREG_PC, gen_im32(dest)); + gen_op_mov32(QREG_T0, gen_im32((long)tb + n)); + gen_op_exit_tb(); + } else { + gen_jmp(s, gen_im32(dest)); + gen_op_mov32(QREG_T0, gen_im32(0)); + gen_op_exit_tb(); + } + s->is_jmp = DISAS_TB_JUMP; +} + +DISAS_INSN(undef_mac) +{ + gen_exception(s, s->pc - 2, EXCP_LINEA); +} + +DISAS_INSN(undef_fpu) +{ + gen_exception(s, s->pc - 2, EXCP_LINEF); +} + +DISAS_INSN(undef) +{ + gen_exception(s, s->pc - 2, EXCP_UNSUPPORTED); + cpu_abort(cpu_single_env, "Illegal instruction: %04x @ %08x", + insn, s->pc - 2); +} + +DISAS_INSN(mulw) +{ + int reg; + int tmp; + int src; + int sign; + + sign = (insn & 0x100) != 0; + reg = DREG(insn, 9); + tmp = gen_new_qreg(QMODE_I32); + if (sign) + gen_op_ext16s32(tmp, reg); + else + gen_op_ext16u32(tmp, reg); + src = gen_ea(s, insn, OS_WORD, sign ? -1 : 0, NULL); + gen_op_mul32(tmp, tmp, src); + gen_op_mov32(reg, tmp); + /* Unlike m68k, coldfire always clears the overflow bit. */ + gen_logic_cc(s, tmp); +} + +DISAS_INSN(divw) +{ + int reg; + int tmp; + int src; + int sign; + + sign = (insn & 0x100) != 0; + reg = DREG(insn, 9); + if (sign) { + gen_op_ext16s32(QREG_DIV1, reg); + } else { + gen_op_ext16u32(QREG_DIV1, reg); + } + src = gen_ea(s, insn, OS_WORD, sign ? -1 : 0, NULL); + gen_op_mov32(QREG_DIV2, src); + if (sign) { + gen_op_divs(1); + } else { + gen_op_divu(1); + } + + tmp = gen_new_qreg(QMODE_I32); + src = gen_new_qreg(QMODE_I32); + gen_op_ext16u32(tmp, QREG_DIV1); + gen_op_shl32(src, QREG_DIV2, gen_im32(16)); + gen_op_or32(reg, tmp, src); + gen_op_flags_set(); + s->cc_op = CC_OP_FLAGS; +} + +DISAS_INSN(divl) +{ + int num; + int den; + int reg; + uint16_t ext; + + ext = lduw(s->pc); + s->pc += 2; + if (ext & 0x87f8) { + gen_exception(s, s->pc - 4, EXCP_UNSUPPORTED); + return; + } + num = DREG(ext, 12); + reg = DREG(ext, 0); + gen_op_mov32(QREG_DIV1, num); + den = gen_ea(s, insn, OS_LONG, 0, NULL); + gen_op_mov32(QREG_DIV2, den); + if (ext & 0x0800) { + gen_op_divs(0); + } else { + gen_op_divu(0); + } + if (num == reg) { + /* div */ + gen_op_mov32 (reg, QREG_DIV1); + } else { + /* rem */ + gen_op_mov32 (reg, QREG_DIV2); + } + gen_op_flags_set(); + s->cc_op = CC_OP_FLAGS; +} + +DISAS_INSN(addsub) +{ + int reg; + int dest; + int src; + int tmp; + int addr; + int add; + + add = (insn & 0x4000) != 0; + reg = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + if (insn & 0x100) { + tmp = gen_ea(s, insn, OS_LONG, 0, &addr); + src = reg; + } else { + tmp = reg; + src = gen_ea(s, insn, OS_LONG, 0, NULL); + } + if (add) { + gen_op_add32(dest, tmp, src); + gen_op_update_xflag_lt(dest, src); + s->cc_op = CC_OP_ADD; + } else { + gen_op_update_xflag_lt(tmp, src); + gen_op_sub32(dest, tmp, src); + s->cc_op = CC_OP_SUB; + } + gen_op_update_cc_add(dest, src); + if (insn & 0x100) { + gen_ea(s, insn, OS_LONG, dest, &addr); + } else { + gen_op_mov32(reg, dest); + } +} + + +/* Reverse the order of the bits in REG. */ +DISAS_INSN(bitrev) +{ + int val; + int tmp1; + int tmp2; + int reg; + + val = gen_new_qreg(QMODE_I32); + tmp1 = gen_new_qreg(QMODE_I32); + tmp2 = gen_new_qreg(QMODE_I32); + reg = DREG(insn, 0); + gen_op_mov32(val, reg); + /* Reverse bits within each nibble. */ + gen_op_shl32(tmp1, val, gen_im32(3)); + gen_op_and32(tmp1, tmp1, gen_im32(0x88888888)); + gen_op_shl32(tmp2, val, gen_im32(1)); + gen_op_and32(tmp2, tmp2, gen_im32(0x44444444)); + gen_op_or32(tmp1, tmp1, tmp2); + gen_op_shr32(tmp2, val, gen_im32(1)); + gen_op_and32(tmp2, tmp2, gen_im32(0x22222222)); + gen_op_or32(tmp1, tmp1, tmp2); + gen_op_shr32(tmp2, val, gen_im32(3)); + gen_op_and32(tmp2, tmp2, gen_im32(0x11111111)); + gen_op_or32(tmp1, tmp1, tmp2); + /* Reverse nibbles withing bytes. */ + gen_op_shl32(val, tmp1, gen_im32(4)); + gen_op_and32(val, val, gen_im32(0xf0f0f0f0)); + gen_op_shr32(tmp2, tmp1, gen_im32(4)); + gen_op_and32(tmp2, tmp2, gen_im32(0x0f0f0f0f)); + gen_op_or32(val, val, tmp2); + /* Reverse bytes. */ + gen_op_bswap32(reg, val); + gen_op_mov32(reg, val); +} + +DISAS_INSN(bitop_reg) +{ + int opsize; + int op; + int src1; + int src2; + int tmp; + int addr; + int dest; + + if ((insn & 0x38) != 0) + opsize = OS_BYTE; + else + opsize = OS_LONG; + op = (insn >> 6) & 3; + src1 = gen_ea(s, insn, opsize, 0, op ? &addr: NULL); + src2 = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + + gen_flush_flags(s); + tmp = gen_new_qreg(QMODE_I32); + if (opsize == OS_BYTE) + gen_op_and32(tmp, src2, gen_im32(7)); + else + gen_op_and32(tmp, src2, gen_im32(31)); + src2 = tmp; + tmp = gen_new_qreg(QMODE_I32); + gen_op_shl32(tmp, gen_im32(1), src2); + + gen_op_btest(src1, tmp); + switch (op) { + case 1: /* bchg */ + gen_op_xor32(dest, src1, tmp); + break; + case 2: /* bclr */ + gen_op_not32(tmp, tmp); + gen_op_and32(dest, src1, tmp); + break; + case 3: /* bset */ + gen_op_or32(dest, src1, tmp); + break; + default: /* btst */ + break; + } + if (op) + gen_ea(s, insn, opsize, dest, &addr); +} + +DISAS_INSN(sats) +{ + int reg; + int tmp; + int l1; + + reg = DREG(insn, 0); + tmp = gen_new_qreg(QMODE_I32); + gen_flush_flags(s); + gen_op_and32(tmp, QREG_CC_DEST, gen_im32(CCF_V)); + l1 = gen_new_label(); + gen_op_jmp_z32(tmp, l1); + tmp = gen_new_qreg(QMODE_I32); + gen_op_shr32(tmp, reg, gen_im32(31)); + gen_op_xor32(tmp, tmp, gen_im32(0x80000000)); + gen_op_mov32(reg, tmp); + gen_set_label(l1); + gen_logic_cc(s, tmp); +} + +static void gen_push(int val) +{ + int tmp; + + tmp = gen_new_qreg(QMODE_I32); + gen_op_sub32(tmp, QREG_SP, gen_im32(4)); + gen_store(OS_LONG, tmp, val); + gen_op_mov32(QREG_SP, tmp); +} + +DISAS_INSN(movem) +{ + int addr; + int i; + uint16_t mask; + int reg; + int tmp; + int is_load; + + mask = lduw(s->pc); + s->pc += 2; + tmp = gen_lea(s, insn, OS_LONG); + addr = gen_new_qreg(QMODE_I32); + gen_op_mov32(addr, tmp); + is_load = ((insn & 0x0400) != 0); + for (i = 0; i < 16; i++, mask >>= 1) { + if (mask & 1) { + if (i < 8) + reg = DREG(i, 0); + else + reg = AREG(i, 0); + if (is_load) { + tmp = gen_load(OS_LONG, addr, 0); + gen_op_mov32(reg, tmp); + } else { + gen_store(OS_LONG, addr, reg); + } + if (mask != 1) + gen_op_add32(addr, addr, gen_im32(4)); + } + } +} + +DISAS_INSN(bitop_im) +{ + int opsize; + int op; + int src1; + uint32_t mask; + int bitnum; + int tmp; + int addr; + int dest; + + if ((insn & 0x38) != 0) + opsize = OS_BYTE; + else + opsize = OS_LONG; + op = (insn >> 6) & 3; + + bitnum = lduw(s->pc); + s->pc += 2; + if (bitnum & 0xff00) { + disas_undef(s, insn); + return; + } + + src1 = gen_ea(s, insn, opsize, 0, op ? &addr: NULL); + + gen_flush_flags(s); + tmp = gen_new_qreg(QMODE_I32); + if (opsize == OS_BYTE) + bitnum &= 7; + else + bitnum &= 31; + mask = 1 << bitnum; + + gen_op_btest(src1, gen_im32(mask)); + if (op) + dest = gen_new_qreg(QMODE_I32); + else + dest = -1; + + switch (op) { + case 1: /* bchg */ + gen_op_xor32(dest, src1, gen_im32(mask)); + break; + case 2: /* bclr */ + gen_op_and32(dest, src1, gen_im32(~mask)); + break; + case 3: /* bset */ + gen_op_or32(dest, src1, gen_im32(mask)); + break; + default: /* btst */ + break; + } + if (op) + gen_ea(s, insn, opsize, dest, &addr); +} + +DISAS_INSN(arith_im) +{ + int op; + int src1; + int dest; + int src2; + int addr; + + op = (insn >> 9) & 7; + src1 = gen_ea(s, insn, OS_LONG, 0, (op == 6) ? NULL : &addr); + src2 = gen_im32(read_im32(s)); + dest = gen_new_qreg(QMODE_I32); + switch (op) { + case 0: /* ori */ + gen_op_or32(dest, src1, src2); + gen_logic_cc(s, dest); + break; + case 1: /* andi */ + gen_op_and32(dest, src1, src2); + gen_logic_cc(s, dest); + break; + case 2: /* subi */ + gen_op_mov32(dest, src1); + gen_op_update_xflag_lt(dest, src2); + gen_op_sub32(dest, dest, src2); + gen_op_update_cc_add(dest, src2); + s->cc_op = CC_OP_SUB; + break; + case 3: /* addi */ + gen_op_mov32(dest, src1); + gen_op_add32(dest, dest, src2); + gen_op_update_cc_add(dest, src2); + gen_op_update_xflag_lt(dest, src2); + s->cc_op = CC_OP_ADD; + break; + case 5: /* eori */ + gen_op_xor32(dest, src1, src2); + gen_logic_cc(s, dest); + break; + case 6: /* cmpi */ + gen_op_mov32(dest, src1); + gen_op_sub32(dest, dest, src2); + gen_op_update_cc_add(dest, src2); + s->cc_op = CC_OP_SUB; + break; + default: + abort(); + } + if (op != 6) { + gen_ea(s, insn, OS_LONG, dest, &addr); + } +} + +DISAS_INSN(byterev) +{ + int reg; + + reg = DREG(insn, 0); + gen_op_bswap32(reg, reg); +} + +DISAS_INSN(move) +{ + int src; + int dest; + int op; + int opsize; + + switch (insn >> 12) { + case 1: /* move.b */ + opsize = OS_BYTE; + break; + case 2: /* move.l */ + opsize = OS_LONG; + break; + case 3: /* move.w */ + opsize = OS_WORD; + break; + default: + abort(); + } + src = gen_ea(s, insn, opsize, -1, NULL); + op = (insn >> 6) & 7; + if (op == 1) { + /* movea */ + /* The value will already have been sign extended. */ + dest = AREG(insn, 9); + gen_op_mov32(dest, src); + } else { + /* normal move */ + uint16_t dest_ea; + dest_ea = ((insn >> 9) & 7) | (op << 3); + gen_ea(s, dest_ea, opsize, src, NULL); + /* This will be correct because loads sign extend. */ + gen_logic_cc(s, src); + } +} + +DISAS_INSN(negx) +{ + int reg; + int dest; + int tmp; + + gen_flush_flags(s); + reg = DREG(insn, 0); + dest = gen_new_qreg(QMODE_I32); + gen_op_mov32 (dest, gen_im32(0)); + gen_op_subx_cc(dest, reg); + /* !Z is sticky. */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_mov32 (tmp, QREG_CC_DEST); + gen_op_update_cc_add(dest, reg); + gen_op_mov32(reg, dest); + s->cc_op = CC_OP_DYNAMIC; + gen_flush_flags(s); + gen_op_or32(tmp, tmp, gen_im32(~CCF_Z)); + gen_op_and32(QREG_CC_DEST, QREG_CC_DEST, tmp); + s->cc_op = CC_OP_FLAGS; +} + +DISAS_INSN(lea) +{ + int reg; + int tmp; + + reg = AREG(insn, 9); + tmp = gen_lea(s, insn, OS_LONG); + gen_op_mov32(reg, tmp); +} + +DISAS_INSN(clr) +{ + int opsize; + + switch ((insn >> 6) & 3) { + case 0: /* clr.b */ + opsize = OS_BYTE; + break; + case 1: /* clr.w */ + opsize = OS_WORD; + break; + case 2: /* clr.l */ + opsize = OS_LONG; + break; + default: + abort(); + } + gen_ea (s, insn, opsize, gen_im32(0), NULL); + gen_logic_cc(s, gen_im32(0)); +} + +DISAS_INSN(move_from_ccr) +{ + int reg; + int dest; + + gen_flush_flags(s); + dest = gen_new_qreg(QMODE_I32); + gen_op_get_xflag(dest); + gen_op_shl32(dest, dest, gen_im32(4)); + gen_op_or32(dest, dest, QREG_CC_DEST); + reg = DREG(insn, 0); + gen_partset_reg(OS_WORD, reg, dest); +} + +DISAS_INSN(neg) +{ + int reg; + int src1; + + reg = DREG(insn, 0); + src1 = gen_new_qreg(QMODE_I32); + gen_op_mov32(src1, reg); + gen_op_neg32(reg, src1); + s->cc_op = CC_OP_SUB; + gen_op_update_cc_add(reg, src1); + gen_op_update_xflag_lt(gen_im32(0), src1); + s->cc_op = CC_OP_SUB; +} + +DISAS_INSN(move_to_ccr) +{ + int src1; + int reg; + + s->cc_op = CC_OP_FLAGS; + if ((insn & 0x38) == 0) + { + src1 = gen_new_qreg(QMODE_I32); + reg = DREG(insn, 0); + gen_op_and32(src1, reg, gen_im32(0xf)); + gen_op_logic_cc(src1); + gen_op_shr32(src1, reg, gen_im32(4)); + gen_op_and32(src1, src1, gen_im32(1)); + gen_op_update_xflag_tst(src1); + } + else if ((insn & 0x3f) != 0x3c) + { + uint8_t val; + val = ldsb(s->pc); + s->pc += 2; + gen_op_logic_cc(gen_im32(val & 0xf)); + gen_op_update_xflag_tst(gen_im32((val & 0x10) >> 4)); + } + else + disas_undef(s, insn); +} + +DISAS_INSN(not) +{ + int reg; + + reg = DREG(insn, 0); + gen_op_not32(reg, reg); + gen_logic_cc(s, reg); +} + +DISAS_INSN(swap) +{ + int dest; + int src1; + int src2; + int reg; + + dest = gen_new_qreg(QMODE_I32); + src1 = gen_new_qreg(QMODE_I32); + src2 = gen_new_qreg(QMODE_I32); + reg = DREG(insn, 0); + gen_op_shl32(src1, reg, gen_im32(16)); + gen_op_shr32(src2, reg, gen_im32(16)); + gen_op_or32(dest, src1, src2); + gen_op_mov32(reg, dest); + gen_logic_cc(s, dest); +} + +DISAS_INSN(pea) +{ + int tmp; + + tmp = gen_lea(s, insn, OS_LONG); + gen_push(tmp); +} + +DISAS_INSN(ext) +{ + int reg; + int op; + int tmp; + + reg = DREG(insn, 0); + op = (insn >> 6) & 7; + tmp = gen_new_qreg(QMODE_I32); + if (op == 3) + gen_op_ext16s32(tmp, reg); + else + gen_op_ext8s32(tmp, reg); + if (op == 2) + gen_partset_reg(OS_WORD, reg, tmp); + else + gen_op_mov32(reg, tmp); + gen_logic_cc(s, tmp); +} + +DISAS_INSN(tst) +{ + int opsize; + int tmp; + + switch ((insn >> 6) & 3) { + case 0: /* tst.b */ + opsize = OS_BYTE; + break; + case 1: /* tst.w */ + opsize = OS_WORD; + break; + case 2: /* tst.l */ + opsize = OS_LONG; + break; + default: + abort(); + } + tmp = gen_ea(s, insn, opsize, -1, NULL); + gen_logic_cc(s, tmp); +} + +DISAS_INSN(pulse) +{ + /* Implemented as a NOP. */ +} + +DISAS_INSN(illegal) +{ + gen_exception(s, s->pc - 2, EXCP_ILLEGAL); +} + +/* ??? This should be atomic. */ +DISAS_INSN(tas) +{ + int dest; + int src1; + int addr; + + dest = gen_new_qreg(QMODE_I32); + src1 = gen_ea(s, insn, OS_BYTE, -1, &addr); + gen_logic_cc(s, src1); + gen_op_or32(dest, src1, gen_im32(0x80)); + gen_ea(s, insn, OS_BYTE, dest, &addr); +} + +DISAS_INSN(mull) +{ + uint16_t ext; + int reg; + int src1; + int dest; + + /* The upper 32 bits of the product are discarded, so + muls.l and mulu.l are functionally equivalent. */ + ext = lduw(s->pc); + s->pc += 2; + if (ext & 0x87ff) { + gen_exception(s, s->pc - 4, EXCP_UNSUPPORTED); + return; + } + reg = DREG(ext, 12); + src1 = gen_ea(s, insn, OS_LONG, 0, NULL); + dest = gen_new_qreg(QMODE_I32); + gen_op_mul32(dest, src1, reg); + gen_op_mov32(reg, dest); + /* Unlike m68k, coldfire always clears the overflow bit. */ + gen_logic_cc(s, dest); +} + +DISAS_INSN(link) +{ + int16_t offset; + int reg; + int tmp; + + offset = ldsw(s->pc); + s->pc += 2; + reg = AREG(insn, 0); + tmp = gen_new_qreg(QMODE_I32); + gen_op_sub32(tmp, QREG_SP, gen_im32(4)); + gen_store(OS_LONG, tmp, reg); + if (reg != QREG_SP) + gen_op_mov32(reg, tmp); + gen_op_add32(QREG_SP, tmp, gen_im32(offset)); +} + +DISAS_INSN(unlk) +{ + int src; + int reg; + int tmp; + + src = gen_new_qreg(QMODE_I32); + reg = AREG(insn, 0); + gen_op_mov32(src, reg); + tmp = gen_load(OS_LONG, src, 0); + gen_op_mov32(reg, tmp); + gen_op_add32(QREG_SP, src, gen_im32(4)); +} + +DISAS_INSN(nop) +{ +} + +DISAS_INSN(rts) +{ + int tmp; + + tmp = gen_load(OS_LONG, QREG_SP, 0); + gen_op_add32(QREG_SP, QREG_SP, gen_im32(4)); + gen_jmp(s, tmp); +} + +DISAS_INSN(jump) +{ + int tmp; + + /* Load the target address first to ensure correct exception + behavior. */ + tmp = gen_lea(s, insn, OS_LONG); + if ((insn & 0x40) == 0) { + /* jsr */ + gen_push(gen_im32(s->pc)); + } + gen_jmp(s, tmp); +} + +DISAS_INSN(addsubq) +{ + int src1; + int src2; + int dest; + int val; + int addr; + + src1 = gen_ea(s, insn, OS_LONG, 0, &addr); + val = (insn >> 9) & 7; + if (val == 0) + val = 8; + src2 = gen_im32(val); + dest = gen_new_qreg(QMODE_I32); + gen_op_mov32(dest, src1); + if ((insn & 0x38) == 0x08) { + /* Don't update condition codes if the destination is an + address register. */ + if (insn & 0x0100) { + gen_op_sub32(dest, dest, src2); + } else { + gen_op_add32(dest, dest, src2); + } + } else { + if (insn & 0x0100) { + gen_op_update_xflag_lt(dest, src2); + gen_op_sub32(dest, dest, src2); + s->cc_op = CC_OP_SUB; + } else { + gen_op_add32(dest, dest, src2); + gen_op_update_xflag_lt(dest, src2); + s->cc_op = CC_OP_ADD; + } + gen_op_update_cc_add(dest, src2); + } + gen_ea(s, insn, OS_LONG, dest, &addr); +} + +DISAS_INSN(tpf) +{ + switch (insn & 7) { + case 2: /* One extension word. */ + s->pc += 2; + break; + case 3: /* Two extension words. */ + s->pc += 4; + break; + case 4: /* No extension words. */ + break; + default: + disas_undef(s, insn); + } +} + +DISAS_INSN(branch) +{ + int32_t offset; + uint32_t base; + int op; + int l1; + + base = s->pc; + op = (insn >> 8) & 0xf; + offset = (int8_t)insn; + if (offset == 0) { + offset = ldsw(s->pc); + s->pc += 2; + } else if (offset == -1) { + offset = read_im32(s); + } + if (op == 1) { + /* bsr */ + gen_push(gen_im32(s->pc)); + } + gen_flush_cc_op(s); + if (op > 1) { + /* Bcc */ + l1 = gen_new_label(); + gen_jmpcc(s, ((insn >> 8) & 0xf) ^ 1, l1); + gen_jmp_tb(s, 1, base + offset); + gen_set_label(l1); + gen_jmp_tb(s, 0, s->pc); + } else { + /* Unconditional branch. */ + gen_jmp_tb(s, 0, base + offset); + } +} + +DISAS_INSN(moveq) +{ + int tmp; + + tmp = gen_im32((int8_t)insn); + gen_op_mov32(DREG(insn, 9), tmp); + gen_logic_cc(s, tmp); +} + +DISAS_INSN(mvzs) +{ + int opsize; + int src; + int reg; + + if (insn & 0x40) + opsize = OS_WORD; + else + opsize = OS_BYTE; + src = gen_ea(s, insn, opsize, (insn & 0x80) ? 0 : -1, NULL); + reg = DREG(insn, 9); + gen_op_mov32(reg, src); + gen_logic_cc(s, src); +} + +DISAS_INSN(or) +{ + int reg; + int dest; + int src; + int addr; + + reg = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + if (insn & 0x100) { + src = gen_ea(s, insn, OS_LONG, 0, &addr); + gen_op_or32(dest, src, reg); + gen_ea(s, insn, OS_LONG, dest, &addr); + } else { + src = gen_ea(s, insn, OS_LONG, 0, NULL); + gen_op_or32(dest, src, reg); + gen_op_mov32(reg, dest); + } + gen_logic_cc(s, dest); +} + +DISAS_INSN(suba) +{ + int src; + int reg; + + src = gen_ea(s, insn, OS_LONG, 0, NULL); + reg = AREG(insn, 9); + gen_op_sub32(reg, reg, src); +} + +DISAS_INSN(subx) +{ + int reg; + int src; + int dest; + int tmp; + + gen_flush_flags(s); + reg = DREG(insn, 9); + src = DREG(insn, 0); + dest = gen_new_qreg(QMODE_I32); + gen_op_mov32 (dest, reg); + gen_op_subx_cc(dest, src); + /* !Z is sticky. */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_mov32 (tmp, QREG_CC_DEST); + gen_op_update_cc_add(dest, src); + gen_op_mov32(reg, dest); + s->cc_op = CC_OP_DYNAMIC; + gen_flush_flags(s); + gen_op_or32(tmp, tmp, gen_im32(~CCF_Z)); + gen_op_and32(QREG_CC_DEST, QREG_CC_DEST, tmp); + s->cc_op = CC_OP_FLAGS; +} + +DISAS_INSN(mov3q) +{ + int src; + int val; + + val = (insn >> 9) & 7; + if (val == 0) + val = -1; + src = gen_im32(val); + gen_logic_cc(s, src); + gen_ea(s, insn, OS_LONG, src, NULL); +} + +DISAS_INSN(cmp) +{ + int op; + int src; + int reg; + int dest; + int opsize; + + op = (insn >> 6) & 3; + switch (op) { + case 0: /* cmp.b */ + opsize = OS_BYTE; + s->cc_op = CC_OP_CMPB; + break; + case 1: /* cmp.w */ + opsize = OS_WORD; + s->cc_op = CC_OP_CMPW; + break; + case 2: /* cmp.l */ + opsize = OS_LONG; + s->cc_op = CC_OP_SUB; + break; + default: + abort(); + } + src = gen_ea(s, insn, opsize, -1, NULL); + reg = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + gen_op_sub32(dest, reg, src); + gen_op_update_cc_add(dest, src); +} + +DISAS_INSN(cmpa) +{ + int opsize; + int src; + int reg; + int dest; + + if (insn & 0x100) { + opsize = OS_LONG; + } else { + opsize = OS_WORD; + } + src = gen_ea(s, insn, opsize, -1, NULL); + reg = AREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + gen_op_sub32(dest, reg, src); + gen_op_update_cc_add(dest, src); + s->cc_op = CC_OP_SUB; +} + +DISAS_INSN(eor) +{ + int src; + int reg; + int dest; + int addr; + + src = gen_ea(s, insn, OS_LONG, 0, &addr); + reg = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + gen_op_xor32(dest, src, reg); + gen_logic_cc(s, dest); + gen_ea(s, insn, OS_LONG, dest, &addr); +} + +DISAS_INSN(and) +{ + int src; + int reg; + int dest; + int addr; + + reg = DREG(insn, 9); + dest = gen_new_qreg(QMODE_I32); + if (insn & 0x100) { + src = gen_ea(s, insn, OS_LONG, 0, &addr); + gen_op_and32(dest, src, reg); + gen_ea(s, insn, OS_LONG, dest, &addr); + } else { + src = gen_ea(s, insn, OS_LONG, 0, NULL); + gen_op_and32(dest, src, reg); + gen_op_mov32(reg, dest); + } + gen_logic_cc(s, dest); +} + +DISAS_INSN(adda) +{ + int src; + int reg; + + src = gen_ea(s, insn, OS_LONG, 0, NULL); + reg = AREG(insn, 9); + gen_op_add32(reg, reg, src); +} + +DISAS_INSN(addx) +{ + int reg; + int src; + int dest; + int tmp; + + gen_flush_flags(s); + reg = DREG(insn, 9); + src = DREG(insn, 0); + dest = gen_new_qreg(QMODE_I32); + gen_op_mov32 (dest, reg); + gen_op_addx_cc(dest, src); + /* !Z is sticky. */ + tmp = gen_new_qreg(QMODE_I32); + gen_op_mov32 (tmp, QREG_CC_DEST); + gen_op_update_cc_add(dest, src); + gen_op_mov32(reg, dest); + s->cc_op = CC_OP_DYNAMIC; + gen_flush_flags(s); + gen_op_or32(tmp, tmp, gen_im32(~CCF_Z)); + gen_op_and32(QREG_CC_DEST, QREG_CC_DEST, tmp); + s->cc_op = CC_OP_FLAGS; +} + +DISAS_INSN(shift_im) +{ + int reg; + int tmp; + + reg = DREG(insn, 0); + tmp = (insn >> 9) & 7; + if (tmp == 0) + tmp = 8; + if (insn & 0x100) { + gen_op_shl_im_cc(reg, tmp); + s->cc_op = CC_OP_SHL; + } else { + if (insn & 8) { + gen_op_shr_im_cc(reg, tmp); + s->cc_op = CC_OP_SHR; + } else { + gen_op_sar_im_cc(reg, tmp); + s->cc_op = CC_OP_SAR; + } + } +} + +DISAS_INSN(shift_reg) +{ + int reg; + int src; + int tmp; + + reg = DREG(insn, 0); + src = DREG(insn, 9); + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, src, gen_im32(63)); + if (insn & 0x100) { + gen_op_shl_cc(reg, tmp); + s->cc_op = CC_OP_SHL; + } else { + if (insn & 8) { + gen_op_shr_cc(reg, tmp); + s->cc_op = CC_OP_SHR; + } else { + gen_op_sar_cc(reg, tmp); + s->cc_op = CC_OP_SAR; + } + } +} + +DISAS_INSN(ff1) +{ + cpu_abort(NULL, "Unimplemented insn: ff1"); +} + +DISAS_INSN(strldsr) +{ + uint16_t ext; + uint32_t addr; + + addr = s->pc - 2; + ext = lduw(s->pc); + s->pc += 2; + if (ext != 0x46FC) + gen_exception(s, addr, EXCP_UNSUPPORTED); + else + gen_exception(s, addr, EXCP_PRIVILEGE); +} + +DISAS_INSN(move_from_sr) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(move_to_sr) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(move_from_usp) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(move_to_usp) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(halt) +{ + gen_exception(s, s->pc, EXCP_HLT); +} + +DISAS_INSN(stop) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(rte) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(movec) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(intouch) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(cpushl) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(wddata) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(wdebug) +{ + gen_exception(s, s->pc - 2, EXCP_PRIVILEGE); +} + +DISAS_INSN(trap) +{ + gen_exception(s, s->pc - 2, EXCP_TRAP0 + (insn & 0xf)); +} + +/* ??? FP exceptions are not implemented. Most exceptions are deferred until + immediately before the next FP instruction is executed. */ +DISAS_INSN(fpu) +{ + uint16_t ext; + int opmode; + int src; + int dest; + int res; + int round; + int opsize; + + ext = lduw(s->pc); + s->pc += 2; + opmode = ext & 0x7f; + switch ((ext >> 13) & 7) { + case 0: case 2: + break; + case 1: + goto undef; + case 3: /* fmove out */ + src = FREG(ext, 7); + /* fmove */ + /* ??? TODO: Proper behavior on overflow. */ + switch ((ext >> 10) & 7) { + case 0: + opsize = OS_LONG; + res = gen_new_qreg(QMODE_I32); + gen_op_f64_to_i32(res, src); + break; + case 1: + opsize = OS_SINGLE; + res = gen_new_qreg(QMODE_F32); + gen_op_f64_to_f32(res, src); + break; + case 4: + opsize = OS_WORD; + res = gen_new_qreg(QMODE_I32); + gen_op_f64_to_i32(res, src); + break; + case 5: + opsize = OS_DOUBLE; + res = src; + break; + case 6: + opsize = OS_BYTE; + res = gen_new_qreg(QMODE_I32); + gen_op_f64_to_i32(res, src); + break; + default: + goto undef; + } + gen_ea(s, insn, opsize, res, NULL); + return; + case 4: /* fmove to control register. */ + switch ((ext >> 10) & 7) { + case 4: /* FPCR */ + /* Not implemented. Ignore writes. */ + break; + case 1: /* FPIAR */ + case 2: /* FPSR */ + default: + cpu_abort(NULL, "Unimplemented: fmove to control %d", + (ext >> 10) & 7); + } + break; + case 5: /* fmove from control register. */ + switch ((ext >> 10) & 7) { + case 4: /* FPCR */ + /* Not implemented. Always return zero. */ + res = gen_im32(0); + break; + case 1: /* FPIAR */ + case 2: /* FPSR */ + default: + cpu_abort(NULL, "Unimplemented: fmove from control %d", + (ext >> 10) & 7); + goto undef; + } + gen_ea(s, insn, OS_LONG, res, NULL); + break; + case 6: /* fmovem */ + case 7: + { + int addr; + uint16_t mask; + if ((ext & 0x1f00) != 0x1000 || (ext & 0xff) == 0) + goto undef; + src = gen_lea(s, insn, OS_LONG); + addr = gen_new_qreg(QMODE_I32); + gen_op_mov32(addr, src); + mask = 0x80; + dest = QREG_F0; + while (mask) { + if (ext & mask) { + if (ext & (1 << 13)) { + /* store */ + gen_op_stf64(addr, dest); + } else { + /* load */ + gen_op_ldf64(dest, addr); + } + if (ext & (mask - 1)) + gen_op_add32(addr, addr, gen_im32(8)); + } + mask >>= 1; + dest++; + } + } + return; + } + if (ext & (1 << 14)) { + int tmp; + + /* Source effective address. */ + switch ((ext >> 10) & 7) { + case 0: opsize = OS_LONG; break; + case 1: opsize = OS_SINGLE; break; + case 4: opsize = OS_WORD; break; + case 5: opsize = OS_DOUBLE; break; + case 6: opsize = OS_BYTE; break; + default: + goto undef; + } + tmp = gen_ea(s, insn, opsize, -1, NULL); + if (opsize == OS_DOUBLE) { + src = tmp; + } else { + src = gen_new_qreg(QMODE_F64); + switch (opsize) { + case OS_LONG: + case OS_WORD: + case OS_BYTE: + gen_op_i32_to_f64(src, tmp); + break; + case OS_SINGLE: + gen_op_f32_to_f64(src, tmp); + break; + } + } + } else { + /* Source register. */ + src = FREG(ext, 10); + } + dest = FREG(ext, 7); + res = gen_new_qreg(QMODE_F64); + if (opmode != 0x3a) + gen_op_movf64(res, dest); + round = 1; + switch (opmode) { + case 0: case 0x40: case 0x44: /* fmove */ + gen_op_movf64(res, src); + break; + case 1: /* fint */ + gen_op_iround_f64(res, src); + round = 0; + break; + case 3: /* fintrz */ + gen_op_itrunc_f64(res, src); + round = 0; + break; + case 4: case 0x41: case 0x45: /* fsqrt */ + gen_op_sqrtf64(res, src); + break; + case 0x18: case 0x58: case 0x5c: /* fabs */ + gen_op_absf64(res, src); + break; + case 0x1a: case 0x5a: case 0x5e: /* fneg */ + gen_op_chsf64(res, src); + break; + case 0x20: case 0x60: case 0x64: /* fdiv */ + gen_op_divf64(res, res, src); + break; + case 0x22: case 0x62: case 0x66: /* fadd */ + gen_op_addf64(res, res, src); + break; + case 0x23: case 0x63: case 0x67: /* fmul */ + gen_op_mulf64(res, res, src); + break; + case 0x28: case 0x68: case 0x6c: /* fsub */ + gen_op_subf64(res, res, src); + break; + case 0x38: /* fcmp */ + gen_op_sub_cmpf64(res, res, src); + dest = 0; + round = 0; + break; + case 0x3a: /* ftst */ + gen_op_movf64(res, src); + dest = 0; + round = 0; + break; + default: + goto undef; + } + if (round) { + if (opmode & 0x40) { + if ((opmode & 0x4) != 0) + round = 0; + } else if ((s->fpcr & M68K_FPCR_PREC) == 0) { + round = 0; + } + } + if (round) { + int tmp; + + tmp = gen_new_qreg(QMODE_F32); + gen_op_f64_to_f32(tmp, res); + gen_op_f32_to_f64(res, tmp); + } + gen_op_fp_result(res); + if (dest) { + gen_op_movf64(dest, res); + } + return; +undef: + s->pc -= 2; + disas_undef_fpu(s, insn); +} + +DISAS_INSN(fbcc) +{ + uint32_t offset; + uint32_t addr; + int flag; + int zero; + int l1; + + addr = s->pc; + offset = ldsw(s->pc); + s->pc += 2; + if (insn & (1 << 6)) { + offset = (offset << 16) | lduw(s->pc); + s->pc += 2; + } + + l1 = gen_new_label(); + /* TODO: Raise BSUN exception. */ + flag = gen_new_qreg(QMODE_I32); + zero = gen_new_qreg(QMODE_F64); + gen_op_zerof64(zero); + gen_op_compare_quietf64(flag, QREG_FP_RESULT, zero); + /* Jump to l1 if condition is true. */ + switch (insn & 0xf) { + case 0: /* f */ + break; + case 1: /* eq (=0) */ + gen_op_jmp_z32(flag, l1); + break; + case 2: /* ogt (=1) */ + gen_op_sub32(flag, flag, gen_im32(1)); + gen_op_jmp_z32(flag, l1); + break; + case 3: /* oge (=0 or =1) */ + gen_op_jmp_z32(flag, l1); + gen_op_sub32(flag, flag, gen_im32(1)); + gen_op_jmp_z32(flag, l1); + break; + case 4: /* olt (=-1) */ + gen_op_jmp_s32(flag, l1); + break; + case 5: /* ole (=-1 or =0) */ + gen_op_jmp_s32(flag, l1); + gen_op_jmp_z32(flag, l1); + break; + case 6: /* ogl (=-1 or =1) */ + gen_op_jmp_s32(flag, l1); + gen_op_sub32(flag, flag, gen_im32(1)); + gen_op_jmp_z32(flag, l1); + break; + case 7: /* or (=2) */ + gen_op_sub32(flag, flag, gen_im32(2)); + gen_op_jmp_z32(flag, l1); + break; + case 8: /* un (<2) */ + gen_op_sub32(flag, flag, gen_im32(2)); + gen_op_jmp_s32(flag, l1); + break; + case 9: /* ueq (=0 or =2) */ + gen_op_jmp_z32(flag, l1); + gen_op_sub32(flag, flag, gen_im32(2)); + gen_op_jmp_z32(flag, l1); + break; + case 10: /* ugt (>0) */ + /* ??? Add jmp_gtu. */ + gen_op_sub32(flag, flag, gen_im32(1)); + gen_op_jmp_ns32(flag, l1); + break; + case 11: /* uge (>=0) */ + gen_op_jmp_ns32(flag, l1); + break; + case 12: /* ult (=-1 or =2) */ + gen_op_jmp_s32(flag, l1); + gen_op_sub32(flag, flag, gen_im32(2)); + gen_op_jmp_z32(flag, l1); + break; + case 13: /* ule (!=1) */ + gen_op_sub32(flag, flag, gen_im32(1)); + gen_op_jmp_nz32(flag, l1); + break; + case 14: /* ne (!=0) */ + gen_op_jmp_nz32(flag, l1); + break; + case 15: /* t */ + gen_op_mov32(flag, gen_im32(1)); + break; + } + gen_jmp_tb(s, 0, s->pc); + gen_set_label(l1); + gen_jmp_tb(s, 1, addr + offset); +} + +static disas_proc opcode_table[65536]; + +static void +register_opcode (disas_proc proc, uint16_t opcode, uint16_t mask) +{ + int i; + int from; + int to; + + /* Sanity check. All set bits must be included in the mask. */ + if (opcode & ~mask) + abort(); + /* This could probably be cleverer. For now just optimize the case where + the top bits are known. */ + /* Find the first zero bit in the mask. */ + i = 0x8000; + while ((i & mask) != 0) + i >>= 1; + /* Iterate over all combinations of this and lower bits. */ + if (i == 0) + i = 1; + else + i <<= 1; + from = opcode & ~(i - 1); + to = from + i; + for (i = from; i < to; i++) + { + if ((i & mask) == opcode) + opcode_table[i] = proc; + } +} + +/* Register m68k opcode handlers. Order is important. + Later insn override earlier ones. */ +static void +register_m68k_insns (m68k_def_t *def) +{ + uint32_t iflags; + + iflags = def->insns; +#define INSN(name, opcode, mask, isa) \ + if (iflags & M68K_INSN_##isa) \ + register_opcode(disas_##name, 0x##opcode, 0x##mask) + INSN(undef, 0000, 0000, CF_A); + INSN(arith_im, 0080, fff8, CF_A); + INSN(bitrev, 00c0, fff8, CF_C); + INSN(bitop_reg, 0100, f1c0, CF_A); + INSN(bitop_reg, 0140, f1c0, CF_A); + INSN(bitop_reg, 0180, f1c0, CF_A); + INSN(bitop_reg, 01c0, f1c0, CF_A); + INSN(arith_im, 0280, fff8, CF_A); + INSN(byterev, 02c0, fff8, CF_A); + INSN(arith_im, 0480, fff8, CF_A); + INSN(ff1, 04c0, fff8, CF_C); + INSN(arith_im, 0680, fff8, CF_A); + INSN(bitop_im, 0800, ffc0, CF_A); + INSN(bitop_im, 0840, ffc0, CF_A); + INSN(bitop_im, 0880, ffc0, CF_A); + INSN(bitop_im, 08c0, ffc0, CF_A); + INSN(arith_im, 0a80, fff8, CF_A); + INSN(arith_im, 0c00, ff38, CF_A); + INSN(move, 1000, f000, CF_A); + INSN(move, 2000, f000, CF_A); + INSN(move, 3000, f000, CF_A); + INSN(strldsr, 40e7, ffff, CF_A); + INSN(negx, 4080, fff8, CF_A); + INSN(move_from_sr, 40c0, fff8, CF_A); + INSN(lea, 41c0, f1c0, CF_A); + INSN(clr, 4200, ff00, CF_A); + INSN(undef, 42c0, ffc0, CF_A); + INSN(move_from_ccr, 42c0, fff8, CF_A); + INSN(neg, 4480, fff8, CF_A); + INSN(move_to_ccr, 44c0, ffc0, CF_A); + INSN(not, 4680, fff8, CF_A); + INSN(move_to_sr, 46c0, ffc0, CF_A); + INSN(pea, 4840, ffc0, CF_A); + INSN(swap, 4840, fff8, CF_A); + INSN(movem, 48c0, fbc0, CF_A); + INSN(ext, 4880, fff8, CF_A); + INSN(ext, 48c0, fff8, CF_A); + INSN(ext, 49c0, fff8, CF_A); + INSN(tst, 4a00, ff00, CF_A); + INSN(tas, 4ac0, ffc0, CF_B); + INSN(halt, 4ac8, ffff, CF_A); + INSN(pulse, 4acc, ffff, CF_A); + INSN(illegal, 4afc, ffff, CF_A); + INSN(mull, 4c00, ffc0, CF_A); + INSN(divl, 4c40, ffc0, CF_A); + INSN(sats, 4c80, fff8, CF_B); + INSN(trap, 4e40, fff0, CF_A); + INSN(link, 4e50, fff8, CF_A); + INSN(unlk, 4e58, fff8, CF_A); + INSN(move_to_usp, 4e60, fff8, CF_B); + INSN(move_from_usp, 4e68, fff8, CF_B); + INSN(nop, 4e71, ffff, CF_A); + INSN(stop, 4e72, ffff, CF_A); + INSN(rte, 4e73, ffff, CF_A); + INSN(rts, 4e75, ffff, CF_A); + INSN(movec, 4e7b, ffff, CF_A); + INSN(jump, 4e80, ffc0, CF_A); + INSN(jump, 4ec0, ffc0, CF_A); + INSN(addsubq, 5180, f1c0, CF_A); + INSN(scc, 50c0, f0f8, CF_A); + INSN(addsubq, 5080, f1c0, CF_A); + INSN(tpf, 51f8, fff8, CF_A); + INSN(branch, 6000, f000, CF_A); + INSN(moveq, 7000, f100, CF_A); + INSN(mvzs, 7100, f100, CF_B); + INSN(or, 8000, f000, CF_A); + INSN(divw, 80c0, f0c0, CF_A); + INSN(addsub, 9000, f000, CF_A); + INSN(subx, 9180, f1f8, CF_A); + INSN(suba, 91c0, f1c0, CF_A); + INSN(undef_mac, a000, f000, CF_A); + INSN(mov3q, a140, f1c0, CF_B); + INSN(cmp, b000, f1c0, CF_B); /* cmp.b */ + INSN(cmp, b040, f1c0, CF_B); /* cmp.w */ + INSN(cmpa, b0c0, f1c0, CF_B); /* cmpa.w */ + INSN(cmp, b080, f1c0, CF_A); + INSN(cmpa, b1c0, f1c0, CF_A); + INSN(eor, b180, f1c0, CF_A); + INSN(and, c000, f000, CF_A); + INSN(mulw, c0c0, f0c0, CF_A); + INSN(addsub, d000, f000, CF_A); + INSN(addx, d180, f1f8, CF_A); + INSN(adda, d1c0, f1c0, CF_A); + INSN(shift_im, e080, f0f0, CF_A); + INSN(shift_reg, e0a0, f0f0, CF_A); + INSN(undef_fpu, f000, f000, CF_A); + INSN(fpu, f200, ffc0, CF_FPU); + INSN(fbcc, f280, ffc0, CF_FPU); + INSN(intouch, f340, ffc0, CF_A); + INSN(cpushl, f428, ff38, CF_A); + INSN(wddata, fb00, ff00, CF_A); + INSN(wdebug, fbc0, ffc0, CF_A); +#undef INSN +} + +/* ??? Some of this implementation is not exception safe. We should always + write back the result to memory before setting the condition codes. */ +static void disas_m68k_insn(CPUState * env, DisasContext *s) +{ + uint16_t insn; + + insn = lduw(s->pc); + s->pc += 2; + + opcode_table[insn](s, insn); +} + +#if 0 +/* Save the result of a floating point operation. */ +static void expand_op_fp_result(qOP *qop) +{ + gen_op_movf64(QREG_FP_RESULT, qop->args[0]); +} + +/* Dummy op to indicate that the flags have been set. */ +static void expand_op_flags_set(qOP *qop) +{ +} + +/* Convert the confition codes into CC_OP_FLAGS format. */ +static void expand_op_flush_flags(qOP *qop) +{ + int cc_opreg; + + if (qop->args[0] == CC_OP_DYNAMIC) + cc_opreg = QREG_CC_OP; + else + cc_opreg = gen_im32(qop->args[0]); + gen_op_helper32(QREG_NULL, cc_opreg, HELPER_flush_flags); +} + +/* Set CC_DEST after a logical or direct flag setting operation. */ +static void expand_op_logic_cc(qOP *qop) +{ + gen_op_mov32(QREG_CC_DEST, qop->args[0]); +} + +/* Set CC_SRC and CC_DEST after an arithmetic operation. */ +static void expand_op_update_cc_add(qOP *qop) +{ + gen_op_mov32(QREG_CC_DEST, qop->args[0]); + gen_op_mov32(QREG_CC_SRC, qop->args[1]); +} + +/* Update the X flag. */ +static void expand_op_update_xflag(qOP *qop) +{ + int arg0; + int arg1; + + arg0 = qop->args[0]; + arg1 = qop->args[1]; + if (arg1 == QREG_NULL) { + /* CC_X = arg0. */ + gen_op_mov32(QREG_CC_X, arg0); + } else { + /* CC_X = arg0 < (unsigned)arg1. */ + gen_op_set_ltu32(QREG_CC_X, arg0, arg1); + } +} + +/* Set arg0 to the contents of the X flag. */ +static void expand_op_get_xflag(qOP *qop) +{ + gen_op_mov32(qop->args[0], QREG_CC_X); +} + +/* Expand a shift by immediate. The ISA only allows shifts by 1-8, so we + already know the shift is within range. */ +static inline void expand_shift_im(qOP *qop, int right, int arith) +{ + int val; + int reg; + int tmp; + int im; + + reg = qop->args[0]; + im = qop->args[1]; + tmp = gen_im32(im); + val = gen_new_qreg(QMODE_I32); + gen_op_mov32(val, reg); + gen_op_mov32(QREG_CC_DEST, val); + gen_op_mov32(QREG_CC_SRC, tmp); + if (right) { + if (arith) { + gen_op_sar32(reg, val, tmp); + } else { + gen_op_shr32(reg, val, tmp); + } + if (im == 1) + tmp = QREG_NULL; + else + tmp = gen_im32(im - 1); + } else { + gen_op_shl32(reg, val, tmp); + tmp = gen_im32(32 - im); + } + if (tmp != QREG_NULL) + gen_op_shr32(val, val, tmp); + gen_op_and32(QREG_CC_X, val, gen_im32(1)); +} + +static void expand_op_shl_im_cc(qOP *qop) +{ + expand_shift_im(qop, 0, 0); +} + +static void expand_op_shr_im_cc(qOP *qop) +{ + expand_shift_im(qop, 1, 0); +} + +static void expand_op_sar_im_cc(qOP *qop) +{ + expand_shift_im(qop, 1, 1); +} + +/* Expand a shift by register. */ +/* ??? This gives incorrect answers for shifts by 0 or >= 32 */ +static inline void expand_shift_reg(qOP *qop, int right, int arith) +{ + int val; + int reg; + int shift; + int tmp; + + reg = qop->args[0]; + shift = qop->args[1]; + val = gen_new_qreg(QMODE_I32); + gen_op_mov32(val, reg); + gen_op_mov32(QREG_CC_DEST, val); + gen_op_mov32(QREG_CC_SRC, shift); + tmp = gen_new_qreg(QMODE_I32); + if (right) { + if (arith) { + gen_op_sar32(reg, val, shift); + } else { + gen_op_shr32(reg, val, shift); + } + gen_op_sub32(tmp, shift, gen_im32(1)); + } else { + gen_op_shl32(reg, val, shift); + gen_op_sub32(tmp, gen_im32(31), shift); + } + gen_op_shl32(val, val, tmp); + gen_op_and32(QREG_CC_X, val, gen_im32(1)); +} + +static void expand_op_shl_cc(qOP *qop) +{ + expand_shift_reg(qop, 0, 0); +} + +static void expand_op_shr_cc(qOP *qop) +{ + expand_shift_reg(qop, 1, 0); +} + +static void expand_op_sar_cc(qOP *qop) +{ + expand_shift_reg(qop, 1, 1); +} + +/* Set the Z flag to (arg0 & arg1) == 0. */ +static void expand_op_btest(qOP *qop) +{ + int tmp; + int l1; + + l1 = gen_new_label(); + tmp = gen_new_qreg(QMODE_I32); + gen_op_and32(tmp, qop->args[0], qop->args[1]); + gen_op_and32(QREG_CC_DEST, QREG_CC_DEST, gen_im32(~(uint32_t)CCF_Z)); + gen_op_jmp_nz32(tmp, l1); + gen_op_or32(QREG_CC_DEST, QREG_CC_DEST, gen_im32(CCF_Z)); + gen_op_label(l1); +} + +/* arg0 += arg1 + CC_X */ +static void expand_op_addx_cc(qOP *qop) +{ + int arg0 = qop->args[0]; + int arg1 = qop->args[1]; + int l1, l2; + + gen_op_add32 (arg0, arg0, arg1); + l1 = gen_new_label(); + l2 = gen_new_label(); + gen_op_jmp_z32(QREG_CC_X, l1); + gen_op_add32(arg0, arg0, gen_im32(1)); + gen_op_mov32(QREG_CC_OP, gen_im32(CC_OP_ADDX)); + gen_op_set_leu32(QREG_CC_X, arg0, arg1); + gen_op_jmp(l2); + gen_set_label(l1); + gen_op_mov32(QREG_CC_OP, gen_im32(CC_OP_ADD)); + gen_op_set_ltu32(QREG_CC_X, arg0, arg1); + gen_set_label(l2); +} + +/* arg0 -= arg1 + CC_X */ +static void expand_op_subx_cc(qOP *qop) +{ + int arg0 = qop->args[0]; + int arg1 = qop->args[1]; + int l1, l2; + + l1 = gen_new_label(); + l2 = gen_new_label(); + gen_op_jmp_z32(QREG_CC_X, l1); + gen_op_set_leu32(QREG_CC_X, arg0, arg1); + gen_op_sub32(arg0, arg0, gen_im32(1)); + gen_op_mov32(QREG_CC_OP, gen_im32(CC_OP_SUBX)); + gen_op_jmp(l2); + gen_set_label(l1); + gen_op_set_ltu32(QREG_CC_X, arg0, arg1); + gen_op_mov32(QREG_CC_OP, gen_im32(CC_OP_SUB)); + gen_set_label(l2); + gen_op_sub32 (arg0, arg0, arg1); +} + +/* Expand target specific ops to generic qops. */ +static void expand_target_qops(void) +{ + qOP *qop; + qOP *next; + int c; + + /* Copy the list of qops, expanding target specific ops as we go. */ + qop = gen_first_qop; + gen_first_qop = NULL; + gen_last_qop = NULL; + for (; qop; qop = next) { + c = qop->opcode; + next = qop->next; + if (c < FIRST_TARGET_OP) { + qop->prev = gen_last_qop; + qop->next = NULL; + if (gen_last_qop) + gen_last_qop->next = qop; + else + gen_first_qop = qop; + gen_last_qop = qop; + continue; + } + switch (c) { +#define DEF(name, nargs, barrier) \ + case INDEX_op_##name: \ + expand_op_##name(qop); \ + break; +#include "qop-target.def" +#undef DEF + default: + cpu_abort(NULL, "Unexpanded target qop"); + } + } +} + +/* ??? Implement this. */ +static void +optimize_flags(void) +{ +} +#endif + +/* generate intermediate code for basic block 'tb'. */ +int gen_intermediate_code_internal(CPUState *env, TranslationBlock *tb, + int search_pc) +{ + DisasContext dc1, *dc = &dc1; + uint16_t *gen_opc_end; + int j, lj; + target_ulong pc_start; + int pc_offset; + int last_cc_op; + + /* generate intermediate code */ + pc_start = tb->pc; + + dc->tb = tb; + + gen_opc_ptr = gen_opc_buf; + gen_opc_end = gen_opc_buf + OPC_MAX_SIZE; + gen_opparam_ptr = gen_opparam_buf; + + dc->is_jmp = DISAS_NEXT; + dc->pc = pc_start; + dc->cc_op = CC_OP_DYNAMIC; + dc->singlestep_enabled = env->singlestep_enabled; + dc->fpcr = env->fpcr; + nb_gen_labels = 0; + lj = -1; + do { + free_qreg = 0; + pc_offset = dc->pc - pc_start; + gen_throws_exception = NULL; + if (env->nb_breakpoints > 0) { + for(j = 0; j < env->nb_breakpoints; j++) { + if (env->breakpoints[j] == dc->pc) { + gen_exception(dc, dc->pc, EXCP_DEBUG); + dc->is_jmp = DISAS_JUMP; + break; + } + } + if (dc->is_jmp) + break; + } + if (search_pc) { + j = gen_opc_ptr - gen_opc_buf; + if (lj < j) { + lj++; + while (lj < j) + gen_opc_instr_start[lj++] = 0; + } + gen_opc_pc[lj] = dc->pc; + gen_opc_instr_start[lj] = 1; + } + last_cc_op = dc->cc_op; + disas_m68k_insn(env, dc); + } while (!dc->is_jmp && gen_opc_ptr < gen_opc_end && + !env->singlestep_enabled && + (pc_offset) < (TARGET_PAGE_SIZE - 32)); + + if (__builtin_expect(env->singlestep_enabled, 0)) { + /* Make sure the pc is updated, and raise a debug exception. */ + if (!dc->is_jmp) { + gen_flush_cc_op(dc); + gen_op_mov32(QREG_PC, gen_im32((long)dc->pc)); + } + gen_op_raise_exception(EXCP_DEBUG); + } else { + switch(dc->is_jmp) { + case DISAS_NEXT: + gen_flush_cc_op(dc); + gen_jmp_tb(dc, 0, dc->pc); + break; + default: + case DISAS_JUMP: + case DISAS_UPDATE: + gen_flush_cc_op(dc); + /* indicate that the hash table must be used to find the next TB */ + gen_op_mov32(QREG_T0, gen_im32(0)); + gen_op_exit_tb(); + break; + case DISAS_TB_JUMP: + /* nothing more to generate */ + break; + } + } + *gen_opc_ptr = INDEX_op_end; + +#ifdef DEBUG_DISAS + if (loglevel & CPU_LOG_TB_IN_ASM) { + fprintf(logfile, "----------------\n"); + fprintf(logfile, "IN: %s\n", lookup_symbol(pc_start)); + target_disas(logfile, pc_start, dc->pc - pc_start, 0); + fprintf(logfile, "\n"); + if (loglevel & (CPU_LOG_TB_OP)) { + fprintf(logfile, "OP:\n"); + dump_ops(gen_opc_buf, gen_opparam_buf); + fprintf(logfile, "\n"); + } + } +#endif + if (search_pc) { + j = gen_opc_ptr - gen_opc_buf; + lj++; + while (lj <= j) + gen_opc_instr_start[lj++] = 0; + tb->size = 0; + } else { + tb->size = dc->pc - pc_start; + } + + //optimize_flags(); + //expand_target_qops(); + return 0; +} + +int gen_intermediate_code(CPUState *env, TranslationBlock *tb) +{ + return gen_intermediate_code_internal(env, tb, 0); +} + +int gen_intermediate_code_pc(CPUState *env, TranslationBlock *tb) +{ + return gen_intermediate_code_internal(env, tb, 1); +} + +CPUM68KState *cpu_m68k_init(void) +{ + CPUM68KState *env; + + env = malloc(sizeof(CPUM68KState)); + if (!env) + return NULL; + cpu_exec_init(env); + + memset(env, 0, sizeof(CPUM68KState)); + /* ??? FP regs should be initialized to NaN. */ + cpu_single_env = env; + env->cc_op = CC_OP_FLAGS; + return env; +} + +void cpu_m68k_close(CPUM68KState *env) +{ + free(env); +} + +m68k_def_t *m68k_find_by_name(const char *name) +{ + m68k_def_t *def; + + def = m68k_cpu_defs; + while (def->name) + { + if (strcmp(def->name, name) == 0) + return def; + def++; + } + return NULL; +} + +void cpu_m68k_register(CPUM68KState *env, m68k_def_t *def) +{ + register_m68k_insns(def); +} + +void cpu_dump_state(CPUState *env, FILE *f, + int (*cpu_fprintf)(FILE *f, const char *fmt, ...), + int flags) +{ + int i; + uint16_t sr; + CPU_DoubleU u; + for (i = 0; i < 8; i++) + { + u.d = env->fregs[i]; + cpu_fprintf (f, "D%d = %08x A%d = %08x F%d = %08x%08x (%12g)\n", + i, env->dregs[i], i, env->aregs[i], + i, u.l.upper, u.l.lower, u.d); + } + cpu_fprintf (f, "PC = %08x ", env->pc); + sr = env->sr; + cpu_fprintf (f, "SR = %04x %c%c%c%c%c ", sr, (sr & 0x10) ? 'X' : '-', + (sr & CCF_N) ? 'N' : '-', (sr & CCF_Z) ? 'Z' : '-', + (sr & CCF_V) ? 'V' : '-', (sr & CCF_C) ? 'C' : '-'); + cpu_fprintf (f, "FPRESULT = %12g\n", env->fp_result); +} + +/* ??? */ +target_ulong cpu_get_phys_page_debug(CPUState *env, target_ulong addr) +{ + return addr; +} + +#if defined(CONFIG_USER_ONLY) + +int cpu_m68k_handle_mmu_fault (CPUState *env, target_ulong address, int rw, + int is_user, int is_softmmu) +{ + env->exception_index = EXCP_ACCESS; + env->mmu.ar = address; + return 1; +} + +#else + +#error not implemented + +#endif |