diff options
author | Alexander Graf <agraf@suse.de> | 2013-12-20 11:01:50 +0100 |
---|---|---|
committer | Aurelien Jarno <aurelien@aurel32.net> | 2013-12-22 19:15:55 +0100 |
commit | f976b09ea249cccc3fd41c98aaf6512908db0bae (patch) | |
tree | c33b9e5bdf4523145a2e44a938f1ac6e942b7373 /target-ppc | |
parent | e8092f7ae1c71419829879e18ce2dd7249a4d00f (diff) | |
download | qemu-f976b09ea249cccc3fd41c98aaf6512908db0bae.zip qemu-f976b09ea249cccc3fd41c98aaf6512908db0bae.tar.gz qemu-f976b09ea249cccc3fd41c98aaf6512908db0bae.tar.bz2 |
PPC: Fix compilation with TCG debug
The recent VSX patches broken compilation of QEMU when configurated
with --enable-debug, as it was treating "target long" TCG variables
as "i64" which is not true for 32bit targets.
This patch fixes all the places that the compiler has found to use
the correct variable type and if necessary manually cast.
Reported-by: Stefan Weil <sw@weilnetz.de>
Signed-off-by: Alexander Graf <agraf@suse.de>
Reviewed-by: Aurelien Jarno <aurelien@aurel32.net>
Signed-off-by: Aurelien Jarno <aurelien@aurel32.net>
Diffstat (limited to 'target-ppc')
-rw-r--r-- | target-ppc/translate.c | 143 |
1 files changed, 81 insertions, 62 deletions
diff --git a/target-ppc/translate.c b/target-ppc/translate.c index ea58dc9..c5c1108 100644 --- a/target-ppc/translate.c +++ b/target-ppc/translate.c @@ -2567,6 +2567,14 @@ static inline void gen_qemu_ld32u(DisasContext *ctx, TCGv arg1, TCGv arg2) } } +static void gen_qemu_ld32u_i64(DisasContext *ctx, TCGv_i64 val, TCGv addr) +{ + TCGv tmp = tcg_temp_new(); + gen_qemu_ld32u(ctx, tmp, addr); + tcg_gen_extu_tl_i64(val, tmp); + tcg_temp_free(tmp); +} + static inline void gen_qemu_ld32s(DisasContext *ctx, TCGv arg1, TCGv arg2) { if (unlikely(ctx->le_mode)) { @@ -2616,6 +2624,14 @@ static inline void gen_qemu_st32(DisasContext *ctx, TCGv arg1, TCGv arg2) } } +static void gen_qemu_st32_i64(DisasContext *ctx, TCGv_i64 val, TCGv addr) +{ + TCGv tmp = tcg_temp_new(); + tcg_gen_trunc_i64_tl(tmp, val); + gen_qemu_st32(ctx, tmp, addr); + tcg_temp_free(tmp); +} + static inline void gen_qemu_st64(DisasContext *ctx, TCGv_i64 arg1, TCGv arg2) { if (unlikely(ctx->le_mode)) { @@ -7048,13 +7064,14 @@ static void gen_lxvdsx(DisasContext *ctx) EA = tcg_temp_new(); gen_addr_reg_index(ctx, EA); gen_qemu_ld64(ctx, cpu_vsrh(xT(ctx->opcode)), EA); - tcg_gen_mov_tl(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode))); + tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode))); tcg_temp_free(EA); } static void gen_lxvw4x(DisasContext *ctx) { - TCGv EA, tmp; + TCGv EA; + TCGv_i64 tmp; TCGv_i64 xth = cpu_vsrh(xT(ctx->opcode)); TCGv_i64 xtl = cpu_vsrl(xT(ctx->opcode)); if (unlikely(!ctx->vsx_enabled)) { @@ -7063,21 +7080,22 @@ static void gen_lxvw4x(DisasContext *ctx) } gen_set_access_type(ctx, ACCESS_INT); EA = tcg_temp_new(); - tmp = tcg_temp_new(); + tmp = tcg_temp_new_i64(); + gen_addr_reg_index(ctx, EA); - gen_qemu_ld32u(ctx, tmp, EA); + gen_qemu_ld32u_i64(ctx, tmp, EA); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_ld32u(ctx, xth, EA); + gen_qemu_ld32u_i64(ctx, xth, EA); tcg_gen_deposit_i64(xth, xth, tmp, 32, 32); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_ld32u(ctx, tmp, EA); + gen_qemu_ld32u_i64(ctx, tmp, EA); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_ld32u(ctx, xtl, EA); + gen_qemu_ld32u_i64(ctx, xtl, EA); tcg_gen_deposit_i64(xtl, xtl, tmp, 32, 32); tcg_temp_free(EA); - tcg_temp_free(tmp); + tcg_temp_free_i64(tmp); } static void gen_stxsdx(DisasContext *ctx) @@ -7112,7 +7130,8 @@ static void gen_stxvd2x(DisasContext *ctx) static void gen_stxvw4x(DisasContext *ctx) { - TCGv EA, tmp; + TCGv_i64 tmp; + TCGv EA; if (unlikely(!ctx->vsx_enabled)) { gen_exception(ctx, POWERPC_EXCP_VSXU); return; @@ -7120,21 +7139,21 @@ static void gen_stxvw4x(DisasContext *ctx) gen_set_access_type(ctx, ACCESS_INT); EA = tcg_temp_new(); gen_addr_reg_index(ctx, EA); - tmp = tcg_temp_new(); + tmp = tcg_temp_new_i64(); tcg_gen_shri_i64(tmp, cpu_vsrh(xS(ctx->opcode)), 32); - gen_qemu_st32(ctx, tmp, EA); + gen_qemu_st32_i64(ctx, tmp, EA); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_st32(ctx, cpu_vsrh(xS(ctx->opcode)), EA); + gen_qemu_st32_i64(ctx, cpu_vsrh(xS(ctx->opcode)), EA); tcg_gen_shri_i64(tmp, cpu_vsrl(xS(ctx->opcode)), 32); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_st32(ctx, tmp, EA); + gen_qemu_st32_i64(ctx, tmp, EA); tcg_gen_addi_tl(EA, EA, 4); - gen_qemu_st32(ctx, cpu_vsrl(xS(ctx->opcode)), EA); + gen_qemu_st32_i64(ctx, cpu_vsrl(xS(ctx->opcode)), EA); tcg_temp_free(EA); - tcg_temp_free(tmp); + tcg_temp_free_i64(tmp); } static void gen_xxpermdi(DisasContext *ctx) @@ -7171,8 +7190,8 @@ static void glue(gen_, name)(DisasContext * ctx) \ gen_exception(ctx, POWERPC_EXCP_VSXU); \ return; \ } \ - xb = tcg_temp_new(); \ - sgm = tcg_temp_new(); \ + xb = tcg_temp_new_i64(); \ + sgm = tcg_temp_new_i64(); \ tcg_gen_mov_i64(xb, cpu_vsrh(xB(ctx->opcode))); \ tcg_gen_movi_i64(sgm, sgn_mask); \ switch (op) { \ @@ -7189,18 +7208,18 @@ static void glue(gen_, name)(DisasContext * ctx) \ break; \ } \ case OP_CPSGN: { \ - TCGv_i64 xa = tcg_temp_new(); \ + TCGv_i64 xa = tcg_temp_new_i64(); \ tcg_gen_mov_i64(xa, cpu_vsrh(xA(ctx->opcode))); \ tcg_gen_and_i64(xa, xa, sgm); \ tcg_gen_andc_i64(xb, xb, sgm); \ tcg_gen_or_i64(xb, xb, xa); \ - tcg_temp_free(xa); \ + tcg_temp_free_i64(xa); \ break; \ } \ } \ tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xb); \ - tcg_temp_free(xb); \ - tcg_temp_free(sgm); \ + tcg_temp_free_i64(xb); \ + tcg_temp_free_i64(sgm); \ } VSX_SCALAR_MOVE(xsabsdp, OP_ABS, SGN_MASK_DP) @@ -7216,9 +7235,9 @@ static void glue(gen_, name)(DisasContext * ctx) \ gen_exception(ctx, POWERPC_EXCP_VSXU); \ return; \ } \ - xbh = tcg_temp_new(); \ - xbl = tcg_temp_new(); \ - sgm = tcg_temp_new(); \ + xbh = tcg_temp_new_i64(); \ + xbl = tcg_temp_new_i64(); \ + sgm = tcg_temp_new_i64(); \ tcg_gen_mov_i64(xbh, cpu_vsrh(xB(ctx->opcode))); \ tcg_gen_mov_i64(xbl, cpu_vsrl(xB(ctx->opcode))); \ tcg_gen_movi_i64(sgm, sgn_mask); \ @@ -7239,8 +7258,8 @@ static void glue(gen_, name)(DisasContext * ctx) \ break; \ } \ case OP_CPSGN: { \ - TCGv_i64 xah = tcg_temp_new(); \ - TCGv_i64 xal = tcg_temp_new(); \ + TCGv_i64 xah = tcg_temp_new_i64(); \ + TCGv_i64 xal = tcg_temp_new_i64(); \ tcg_gen_mov_i64(xah, cpu_vsrh(xA(ctx->opcode))); \ tcg_gen_mov_i64(xal, cpu_vsrl(xA(ctx->opcode))); \ tcg_gen_and_i64(xah, xah, sgm); \ @@ -7249,16 +7268,16 @@ static void glue(gen_, name)(DisasContext * ctx) \ tcg_gen_andc_i64(xbl, xbl, sgm); \ tcg_gen_or_i64(xbh, xbh, xah); \ tcg_gen_or_i64(xbl, xbl, xal); \ - tcg_temp_free(xah); \ - tcg_temp_free(xal); \ + tcg_temp_free_i64(xah); \ + tcg_temp_free_i64(xal); \ break; \ } \ } \ tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xbh); \ tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), xbl); \ - tcg_temp_free(xbh); \ - tcg_temp_free(xbl); \ - tcg_temp_free(sgm); \ + tcg_temp_free_i64(xbh); \ + tcg_temp_free_i64(xbl); \ + tcg_temp_free_i64(sgm); \ } VSX_VECTOR_MOVE(xvabsdp, OP_ABS, SGN_MASK_DP) @@ -7284,11 +7303,11 @@ static void glue(gen_, name)(DisasContext * ctx) \ cpu_vsrl(xB(ctx->opcode))); \ } -VSX_LOGICAL(xxland, tcg_gen_and_tl) -VSX_LOGICAL(xxlandc, tcg_gen_andc_tl) -VSX_LOGICAL(xxlor, tcg_gen_or_tl) -VSX_LOGICAL(xxlxor, tcg_gen_xor_tl) -VSX_LOGICAL(xxlnor, tcg_gen_nor_tl) +VSX_LOGICAL(xxland, tcg_gen_and_i64) +VSX_LOGICAL(xxlandc, tcg_gen_andc_i64) +VSX_LOGICAL(xxlor, tcg_gen_or_i64) +VSX_LOGICAL(xxlxor, tcg_gen_xor_i64) +VSX_LOGICAL(xxlnor, tcg_gen_nor_i64) #define VSX_XXMRG(name, high) \ static void glue(gen_, name)(DisasContext * ctx) \ @@ -7298,10 +7317,10 @@ static void glue(gen_, name)(DisasContext * ctx) \ gen_exception(ctx, POWERPC_EXCP_VSXU); \ return; \ } \ - a0 = tcg_temp_new(); \ - a1 = tcg_temp_new(); \ - b0 = tcg_temp_new(); \ - b1 = tcg_temp_new(); \ + a0 = tcg_temp_new_i64(); \ + a1 = tcg_temp_new_i64(); \ + b0 = tcg_temp_new_i64(); \ + b1 = tcg_temp_new_i64(); \ if (high) { \ tcg_gen_mov_i64(a0, cpu_vsrh(xA(ctx->opcode))); \ tcg_gen_mov_i64(a1, cpu_vsrh(xA(ctx->opcode))); \ @@ -7319,10 +7338,10 @@ static void glue(gen_, name)(DisasContext * ctx) \ b0, a0, 32, 32); \ tcg_gen_deposit_i64(cpu_vsrl(xT(ctx->opcode)), \ b1, a1, 32, 32); \ - tcg_temp_free(a0); \ - tcg_temp_free(a1); \ - tcg_temp_free(b0); \ - tcg_temp_free(b1); \ + tcg_temp_free_i64(a0); \ + tcg_temp_free_i64(a1); \ + tcg_temp_free_i64(b0); \ + tcg_temp_free_i64(b1); \ } VSX_XXMRG(xxmrghw, 1) @@ -7335,9 +7354,9 @@ static void gen_xxsel(DisasContext * ctx) gen_exception(ctx, POWERPC_EXCP_VSXU); return; } - a = tcg_temp_new(); - b = tcg_temp_new(); - c = tcg_temp_new(); + a = tcg_temp_new_i64(); + b = tcg_temp_new_i64(); + c = tcg_temp_new_i64(); tcg_gen_mov_i64(a, cpu_vsrh(xA(ctx->opcode))); tcg_gen_mov_i64(b, cpu_vsrh(xB(ctx->opcode))); @@ -7355,9 +7374,9 @@ static void gen_xxsel(DisasContext * ctx) tcg_gen_andc_i64(a, a, c); tcg_gen_or_i64(cpu_vsrl(xT(ctx->opcode)), a, b); - tcg_temp_free(a); - tcg_temp_free(b); - tcg_temp_free(c); + tcg_temp_free_i64(a); + tcg_temp_free_i64(b); + tcg_temp_free_i64(c); } static void gen_xxspltw(DisasContext *ctx) @@ -7372,8 +7391,8 @@ static void gen_xxspltw(DisasContext *ctx) return; } - b = tcg_temp_new(); - b2 = tcg_temp_new(); + b = tcg_temp_new_i64(); + b2 = tcg_temp_new_i64(); if (UIM(ctx->opcode) & 1) { tcg_gen_ext32u_i64(b, vsr); @@ -7385,8 +7404,8 @@ static void gen_xxspltw(DisasContext *ctx) tcg_gen_or_i64(cpu_vsrh(xT(ctx->opcode)), b, b2); tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode))); - tcg_temp_free(b); - tcg_temp_free(b2); + tcg_temp_free_i64(b); + tcg_temp_free_i64(b2); } static void gen_xxsldwi(DisasContext *ctx) @@ -7396,8 +7415,8 @@ static void gen_xxsldwi(DisasContext *ctx) gen_exception(ctx, POWERPC_EXCP_VSXU); return; } - xth = tcg_temp_new(); - xtl = tcg_temp_new(); + xth = tcg_temp_new_i64(); + xtl = tcg_temp_new_i64(); switch (SHW(ctx->opcode)) { case 0: { @@ -7406,7 +7425,7 @@ static void gen_xxsldwi(DisasContext *ctx) break; } case 1: { - TCGv_i64 t0 = tcg_temp_new(); + TCGv_i64 t0 = tcg_temp_new_i64(); tcg_gen_mov_i64(xth, cpu_vsrh(xA(ctx->opcode))); tcg_gen_shli_i64(xth, xth, 32); tcg_gen_mov_i64(t0, cpu_vsrl(xA(ctx->opcode))); @@ -7417,7 +7436,7 @@ static void gen_xxsldwi(DisasContext *ctx) tcg_gen_mov_i64(t0, cpu_vsrh(xB(ctx->opcode))); tcg_gen_shri_i64(t0, t0, 32); tcg_gen_or_i64(xtl, xtl, t0); - tcg_temp_free(t0); + tcg_temp_free_i64(t0); break; } case 2: { @@ -7426,7 +7445,7 @@ static void gen_xxsldwi(DisasContext *ctx) break; } case 3: { - TCGv_i64 t0 = tcg_temp_new(); + TCGv_i64 t0 = tcg_temp_new_i64(); tcg_gen_mov_i64(xth, cpu_vsrl(xA(ctx->opcode))); tcg_gen_shli_i64(xth, xth, 32); tcg_gen_mov_i64(t0, cpu_vsrh(xB(ctx->opcode))); @@ -7437,7 +7456,7 @@ static void gen_xxsldwi(DisasContext *ctx) tcg_gen_mov_i64(t0, cpu_vsrl(xB(ctx->opcode))); tcg_gen_shri_i64(t0, t0, 32); tcg_gen_or_i64(xtl, xtl, t0); - tcg_temp_free(t0); + tcg_temp_free_i64(t0); break; } } @@ -7445,8 +7464,8 @@ static void gen_xxsldwi(DisasContext *ctx) tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xth); tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), xtl); - tcg_temp_free(xth); - tcg_temp_free(xtl); + tcg_temp_free_i64(xth); + tcg_temp_free_i64(xtl); } |