4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
28 # define LOG_DISAS(...) do { } while (0)
32 #include "disas/disas.h"
35 #include "qemu/host-utils.h"
37 /* global register indexes */
38 static TCGv_ptr cpu_env;
40 #include "exec/gen-icount.h"
46 /* Information that (most) every instruction needs to manipulate. */
47 typedef struct DisasContext DisasContext;
48 typedef struct DisasInsn DisasInsn;
49 typedef struct DisasFields DisasFields;
52 struct TranslationBlock *tb;
53 const DisasInsn *insn;
57 bool singlestep_enabled;
61 /* Information carried about a condition to be evaluated. */
68 struct { TCGv_i64 a, b; } s64;
69 struct { TCGv_i32 a, b; } s32;
75 static void gen_op_calc_cc(DisasContext *s);
77 #ifdef DEBUG_INLINE_BRANCHES
78 static uint64_t inline_branch_hit[CC_OP_MAX];
79 static uint64_t inline_branch_miss[CC_OP_MAX];
82 static inline void debug_insn(uint64_t insn)
84 LOG_DISAS("insn: 0x%" PRIx64 "\n", insn);
87 static inline uint64_t pc_to_link_info(DisasContext *s, uint64_t pc)
89 if (!(s->tb->flags & FLAG_MASK_64)) {
90 if (s->tb->flags & FLAG_MASK_32) {
91 return pc | 0x80000000;
97 void cpu_dump_state(CPUS390XState *env, FILE *f, fprintf_function cpu_fprintf,
102 if (env->cc_op > 3) {
103 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %15s\n",
104 env->psw.mask, env->psw.addr, cc_name(env->cc_op));
106 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %02x\n",
107 env->psw.mask, env->psw.addr, env->cc_op);
110 for (i = 0; i < 16; i++) {
111 cpu_fprintf(f, "R%02d=%016" PRIx64, i, env->regs[i]);
113 cpu_fprintf(f, "\n");
119 for (i = 0; i < 16; i++) {
120 cpu_fprintf(f, "F%02d=%016" PRIx64, i, env->fregs[i].ll);
122 cpu_fprintf(f, "\n");
128 #ifndef CONFIG_USER_ONLY
129 for (i = 0; i < 16; i++) {
130 cpu_fprintf(f, "C%02d=%016" PRIx64, i, env->cregs[i]);
132 cpu_fprintf(f, "\n");
139 #ifdef DEBUG_INLINE_BRANCHES
140 for (i = 0; i < CC_OP_MAX; i++) {
141 cpu_fprintf(f, " %15s = %10ld\t%10ld\n", cc_name(i),
142 inline_branch_miss[i], inline_branch_hit[i]);
146 cpu_fprintf(f, "\n");
149 static TCGv_i64 psw_addr;
150 static TCGv_i64 psw_mask;
152 static TCGv_i32 cc_op;
153 static TCGv_i64 cc_src;
154 static TCGv_i64 cc_dst;
155 static TCGv_i64 cc_vr;
157 static char cpu_reg_names[32][4];
158 static TCGv_i64 regs[16];
159 static TCGv_i64 fregs[16];
161 static uint8_t gen_opc_cc_op[OPC_BUF_SIZE];
163 void s390x_translate_init(void)
167 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
168 psw_addr = tcg_global_mem_new_i64(TCG_AREG0,
169 offsetof(CPUS390XState, psw.addr),
171 psw_mask = tcg_global_mem_new_i64(TCG_AREG0,
172 offsetof(CPUS390XState, psw.mask),
175 cc_op = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUS390XState, cc_op),
177 cc_src = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_src),
179 cc_dst = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_dst),
181 cc_vr = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_vr),
184 for (i = 0; i < 16; i++) {
185 snprintf(cpu_reg_names[i], sizeof(cpu_reg_names[0]), "r%d", i);
186 regs[i] = tcg_global_mem_new(TCG_AREG0,
187 offsetof(CPUS390XState, regs[i]),
191 for (i = 0; i < 16; i++) {
192 snprintf(cpu_reg_names[i + 16], sizeof(cpu_reg_names[0]), "f%d", i);
193 fregs[i] = tcg_global_mem_new(TCG_AREG0,
194 offsetof(CPUS390XState, fregs[i].d),
195 cpu_reg_names[i + 16]);
198 /* register helpers */
203 static inline TCGv_i64 load_reg(int reg)
205 TCGv_i64 r = tcg_temp_new_i64();
206 tcg_gen_mov_i64(r, regs[reg]);
210 static inline TCGv_i64 load_freg(int reg)
212 TCGv_i64 r = tcg_temp_new_i64();
213 tcg_gen_mov_i64(r, fregs[reg]);
217 static inline TCGv_i32 load_freg32(int reg)
219 TCGv_i32 r = tcg_temp_new_i32();
220 #if HOST_LONG_BITS == 32
221 tcg_gen_mov_i32(r, TCGV_HIGH(fregs[reg]));
223 tcg_gen_shri_i64(MAKE_TCGV_I64(GET_TCGV_I32(r)), fregs[reg], 32);
228 static inline TCGv_i64 load_freg32_i64(int reg)
230 TCGv_i64 r = tcg_temp_new_i64();
231 tcg_gen_shri_i64(r, fregs[reg], 32);
235 static inline TCGv_i32 load_reg32(int reg)
237 TCGv_i32 r = tcg_temp_new_i32();
238 tcg_gen_trunc_i64_i32(r, regs[reg]);
242 static inline TCGv_i64 load_reg32_i64(int reg)
244 TCGv_i64 r = tcg_temp_new_i64();
245 tcg_gen_ext32s_i64(r, regs[reg]);
249 static inline void store_reg(int reg, TCGv_i64 v)
251 tcg_gen_mov_i64(regs[reg], v);
254 static inline void store_freg(int reg, TCGv_i64 v)
256 tcg_gen_mov_i64(fregs[reg], v);
259 static inline void store_reg32(int reg, TCGv_i32 v)
261 /* 32 bit register writes keep the upper half */
262 #if HOST_LONG_BITS == 32
263 tcg_gen_mov_i32(TCGV_LOW(regs[reg]), v);
265 tcg_gen_deposit_i64(regs[reg], regs[reg],
266 MAKE_TCGV_I64(GET_TCGV_I32(v)), 0, 32);
270 static inline void store_reg32_i64(int reg, TCGv_i64 v)
272 /* 32 bit register writes keep the upper half */
273 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 0, 32);
276 static inline void store_reg32h_i64(int reg, TCGv_i64 v)
278 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 32, 32);
281 static inline void store_freg32(int reg, TCGv_i32 v)
283 /* 32 bit register writes keep the lower half */
284 #if HOST_LONG_BITS == 32
285 tcg_gen_mov_i32(TCGV_HIGH(fregs[reg]), v);
287 tcg_gen_deposit_i64(fregs[reg], fregs[reg],
288 MAKE_TCGV_I64(GET_TCGV_I32(v)), 32, 32);
292 static inline void store_freg32_i64(int reg, TCGv_i64 v)
294 tcg_gen_deposit_i64(fregs[reg], fregs[reg], v, 32, 32);
297 static inline void return_low128(TCGv_i64 dest)
299 tcg_gen_ld_i64(dest, cpu_env, offsetof(CPUS390XState, retxl));
302 static inline void update_psw_addr(DisasContext *s)
305 tcg_gen_movi_i64(psw_addr, s->pc);
308 static inline void potential_page_fault(DisasContext *s)
310 #ifndef CONFIG_USER_ONLY
316 static inline uint64_t ld_code2(CPUS390XState *env, uint64_t pc)
318 return (uint64_t)cpu_lduw_code(env, pc);
321 static inline uint64_t ld_code4(CPUS390XState *env, uint64_t pc)
323 return (uint64_t)(uint32_t)cpu_ldl_code(env, pc);
326 static inline uint64_t ld_code6(CPUS390XState *env, uint64_t pc)
328 return (ld_code2(env, pc) << 32) | ld_code4(env, pc + 2);
331 static inline int get_mem_index(DisasContext *s)
333 switch (s->tb->flags & FLAG_MASK_ASC) {
334 case PSW_ASC_PRIMARY >> 32:
336 case PSW_ASC_SECONDARY >> 32:
338 case PSW_ASC_HOME >> 32:
346 static void gen_exception(int excp)
348 TCGv_i32 tmp = tcg_const_i32(excp);
349 gen_helper_exception(cpu_env, tmp);
350 tcg_temp_free_i32(tmp);
353 static void gen_program_exception(DisasContext *s, int code)
357 /* Remember what pgm exeption this was. */
358 tmp = tcg_const_i32(code);
359 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_code));
360 tcg_temp_free_i32(tmp);
362 tmp = tcg_const_i32(s->next_pc - s->pc);
363 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_ilen));
364 tcg_temp_free_i32(tmp);
366 /* Advance past instruction. */
373 /* Trigger exception. */
374 gen_exception(EXCP_PGM);
377 s->is_jmp = DISAS_EXCP;
380 static inline void gen_illegal_opcode(DisasContext *s)
382 gen_program_exception(s, PGM_SPECIFICATION);
385 static inline void check_privileged(DisasContext *s)
387 if (s->tb->flags & (PSW_MASK_PSTATE >> 32)) {
388 gen_program_exception(s, PGM_PRIVILEGED);
392 static TCGv_i64 get_address(DisasContext *s, int x2, int b2, int d2)
396 /* 31-bitify the immediate part; register contents are dealt with below */
397 if (!(s->tb->flags & FLAG_MASK_64)) {
403 tmp = tcg_const_i64(d2);
404 tcg_gen_add_i64(tmp, tmp, regs[x2]);
409 tcg_gen_add_i64(tmp, tmp, regs[b2]);
413 tmp = tcg_const_i64(d2);
414 tcg_gen_add_i64(tmp, tmp, regs[b2]);
419 tmp = tcg_const_i64(d2);
422 /* 31-bit mode mask if there are values loaded from registers */
423 if (!(s->tb->flags & FLAG_MASK_64) && (x2 || b2)) {
424 tcg_gen_andi_i64(tmp, tmp, 0x7fffffffUL);
430 static inline void gen_op_movi_cc(DisasContext *s, uint32_t val)
432 s->cc_op = CC_OP_CONST0 + val;
435 static void gen_op_update1_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 dst)
437 tcg_gen_discard_i64(cc_src);
438 tcg_gen_mov_i64(cc_dst, dst);
439 tcg_gen_discard_i64(cc_vr);
443 static void gen_op_update1_cc_i32(DisasContext *s, enum cc_op op, TCGv_i32 dst)
445 tcg_gen_discard_i64(cc_src);
446 tcg_gen_extu_i32_i64(cc_dst, dst);
447 tcg_gen_discard_i64(cc_vr);
451 static void gen_op_update2_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
454 tcg_gen_mov_i64(cc_src, src);
455 tcg_gen_mov_i64(cc_dst, dst);
456 tcg_gen_discard_i64(cc_vr);
460 static void gen_op_update2_cc_i32(DisasContext *s, enum cc_op op, TCGv_i32 src,
463 tcg_gen_extu_i32_i64(cc_src, src);
464 tcg_gen_extu_i32_i64(cc_dst, dst);
465 tcg_gen_discard_i64(cc_vr);
469 static void gen_op_update3_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
470 TCGv_i64 dst, TCGv_i64 vr)
472 tcg_gen_mov_i64(cc_src, src);
473 tcg_gen_mov_i64(cc_dst, dst);
474 tcg_gen_mov_i64(cc_vr, vr);
478 static inline void set_cc_nz_u32(DisasContext *s, TCGv_i32 val)
480 gen_op_update1_cc_i32(s, CC_OP_NZ, val);
483 static inline void set_cc_nz_u64(DisasContext *s, TCGv_i64 val)
485 gen_op_update1_cc_i64(s, CC_OP_NZ, val);
488 static inline void gen_set_cc_nz_f32(DisasContext *s, TCGv_i64 val)
490 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, val);
493 static inline void gen_set_cc_nz_f64(DisasContext *s, TCGv_i64 val)
495 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, val);
498 static inline void gen_set_cc_nz_f128(DisasContext *s, TCGv_i64 vh, TCGv_i64 vl)
500 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, vh, vl);
503 static inline void cmp_32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2,
506 gen_op_update2_cc_i32(s, cond, v1, v2);
509 static inline void cmp_64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2,
512 gen_op_update2_cc_i64(s, cond, v1, v2);
515 static inline void cmp_s32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2)
517 cmp_32(s, v1, v2, CC_OP_LTGT_32);
520 static inline void cmp_u32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2)
522 cmp_32(s, v1, v2, CC_OP_LTUGTU_32);
525 static inline void cmp_s32c(DisasContext *s, TCGv_i32 v1, int32_t v2)
527 /* XXX optimize for the constant? put it in s? */
528 TCGv_i32 tmp = tcg_const_i32(v2);
529 cmp_32(s, v1, tmp, CC_OP_LTGT_32);
530 tcg_temp_free_i32(tmp);
533 static inline void cmp_u32c(DisasContext *s, TCGv_i32 v1, uint32_t v2)
535 TCGv_i32 tmp = tcg_const_i32(v2);
536 cmp_32(s, v1, tmp, CC_OP_LTUGTU_32);
537 tcg_temp_free_i32(tmp);
540 static inline void cmp_s64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2)
542 cmp_64(s, v1, v2, CC_OP_LTGT_64);
545 static inline void cmp_u64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2)
547 cmp_64(s, v1, v2, CC_OP_LTUGTU_64);
550 static inline void cmp_s64c(DisasContext *s, TCGv_i64 v1, int64_t v2)
552 TCGv_i64 tmp = tcg_const_i64(v2);
554 tcg_temp_free_i64(tmp);
557 static inline void cmp_u64c(DisasContext *s, TCGv_i64 v1, uint64_t v2)
559 TCGv_i64 tmp = tcg_const_i64(v2);
561 tcg_temp_free_i64(tmp);
564 static inline void set_cc_s32(DisasContext *s, TCGv_i32 val)
566 gen_op_update1_cc_i32(s, CC_OP_LTGT0_32, val);
569 static inline void set_cc_s64(DisasContext *s, TCGv_i64 val)
571 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, val);
574 /* CC value is in env->cc_op */
575 static inline void set_cc_static(DisasContext *s)
577 tcg_gen_discard_i64(cc_src);
578 tcg_gen_discard_i64(cc_dst);
579 tcg_gen_discard_i64(cc_vr);
580 s->cc_op = CC_OP_STATIC;
583 static inline void gen_op_set_cc_op(DisasContext *s)
585 if (s->cc_op != CC_OP_DYNAMIC && s->cc_op != CC_OP_STATIC) {
586 tcg_gen_movi_i32(cc_op, s->cc_op);
590 static inline void gen_update_cc_op(DisasContext *s)
595 /* calculates cc into cc_op */
596 static void gen_op_calc_cc(DisasContext *s)
598 TCGv_i32 local_cc_op = tcg_const_i32(s->cc_op);
599 TCGv_i64 dummy = tcg_const_i64(0);
606 /* s->cc_op is the cc value */
607 tcg_gen_movi_i32(cc_op, s->cc_op - CC_OP_CONST0);
610 /* env->cc_op already is the cc value */
625 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, dummy, cc_dst, dummy);
630 case CC_OP_LTUGTU_32:
631 case CC_OP_LTUGTU_64:
638 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, dummy);
653 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, cc_vr);
656 /* unknown operation - assume 3 arguments and cc_op in env */
657 gen_helper_calc_cc(cc_op, cpu_env, cc_op, cc_src, cc_dst, cc_vr);
663 tcg_temp_free_i32(local_cc_op);
664 tcg_temp_free_i64(dummy);
666 /* We now have cc in cc_op as constant */
670 static inline void decode_rr(DisasContext *s, uint64_t insn, int *r1, int *r2)
674 *r1 = (insn >> 4) & 0xf;
678 static inline TCGv_i64 decode_rx(DisasContext *s, uint64_t insn, int *r1,
679 int *x2, int *b2, int *d2)
683 *r1 = (insn >> 20) & 0xf;
684 *x2 = (insn >> 16) & 0xf;
685 *b2 = (insn >> 12) & 0xf;
688 return get_address(s, *x2, *b2, *d2);
691 static inline void decode_rs(DisasContext *s, uint64_t insn, int *r1, int *r3,
696 *r1 = (insn >> 20) & 0xf;
698 *r3 = (insn >> 16) & 0xf;
699 *b2 = (insn >> 12) & 0xf;
703 static inline TCGv_i64 decode_si(DisasContext *s, uint64_t insn, int *i2,
708 *i2 = (insn >> 16) & 0xff;
709 *b1 = (insn >> 12) & 0xf;
712 return get_address(s, 0, *b1, *d1);
715 static int use_goto_tb(DisasContext *s, uint64_t dest)
717 /* NOTE: we handle the case where the TB spans two pages here */
718 return (((dest & TARGET_PAGE_MASK) == (s->tb->pc & TARGET_PAGE_MASK)
719 || (dest & TARGET_PAGE_MASK) == ((s->pc - 1) & TARGET_PAGE_MASK))
720 && !s->singlestep_enabled
721 && !(s->tb->cflags & CF_LAST_IO));
724 static inline void gen_goto_tb(DisasContext *s, int tb_num, target_ulong pc)
728 if (use_goto_tb(s, pc)) {
729 tcg_gen_goto_tb(tb_num);
730 tcg_gen_movi_i64(psw_addr, pc);
731 tcg_gen_exit_tb((tcg_target_long)s->tb + tb_num);
733 /* jump to another page: currently not optimized */
734 tcg_gen_movi_i64(psw_addr, pc);
739 static inline void account_noninline_branch(DisasContext *s, int cc_op)
741 #ifdef DEBUG_INLINE_BRANCHES
742 inline_branch_miss[cc_op]++;
746 static inline void account_inline_branch(DisasContext *s, int cc_op)
748 #ifdef DEBUG_INLINE_BRANCHES
749 inline_branch_hit[cc_op]++;
753 /* Table of mask values to comparison codes, given a comparison as input.
754 For a true comparison CC=3 will never be set, but we treat this
755 conservatively for possible use when CC=3 indicates overflow. */
756 static const TCGCond ltgt_cond[16] = {
757 TCG_COND_NEVER, TCG_COND_NEVER, /* | | | x */
758 TCG_COND_GT, TCG_COND_NEVER, /* | | GT | x */
759 TCG_COND_LT, TCG_COND_NEVER, /* | LT | | x */
760 TCG_COND_NE, TCG_COND_NEVER, /* | LT | GT | x */
761 TCG_COND_EQ, TCG_COND_NEVER, /* EQ | | | x */
762 TCG_COND_GE, TCG_COND_NEVER, /* EQ | | GT | x */
763 TCG_COND_LE, TCG_COND_NEVER, /* EQ | LT | | x */
764 TCG_COND_ALWAYS, TCG_COND_ALWAYS, /* EQ | LT | GT | x */
767 /* Table of mask values to comparison codes, given a logic op as input.
768 For such, only CC=0 and CC=1 should be possible. */
769 static const TCGCond nz_cond[16] = {
771 TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER,
773 TCG_COND_NE, TCG_COND_NE, TCG_COND_NE, TCG_COND_NE,
775 TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ,
776 /* EQ | NE | x | x */
777 TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS,
780 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
781 details required to generate a TCG comparison. */
782 static void disas_jcc(DisasContext *s, DisasCompare *c, uint32_t mask)
785 enum cc_op old_cc_op = s->cc_op;
787 if (mask == 15 || mask == 0) {
788 c->cond = (mask ? TCG_COND_ALWAYS : TCG_COND_NEVER);
791 c->g1 = c->g2 = true;
796 /* Find the TCG condition for the mask + cc op. */
802 cond = ltgt_cond[mask];
803 if (cond == TCG_COND_NEVER) {
806 account_inline_branch(s, old_cc_op);
809 case CC_OP_LTUGTU_32:
810 case CC_OP_LTUGTU_64:
811 cond = tcg_unsigned_cond(ltgt_cond[mask]);
812 if (cond == TCG_COND_NEVER) {
815 account_inline_branch(s, old_cc_op);
819 cond = nz_cond[mask];
820 if (cond == TCG_COND_NEVER) {
823 account_inline_branch(s, old_cc_op);
838 account_inline_branch(s, old_cc_op);
853 account_inline_branch(s, old_cc_op);
857 switch (mask & 0xa) {
858 case 8: /* src == 0 -> no one bit found */
861 case 2: /* src != 0 -> one bit found */
867 account_inline_branch(s, old_cc_op);
872 /* Calculate cc value. */
877 /* Jump based on CC. We'll load up the real cond below;
878 the assignment here merely avoids a compiler warning. */
879 account_noninline_branch(s, old_cc_op);
880 old_cc_op = CC_OP_STATIC;
881 cond = TCG_COND_NEVER;
885 /* Load up the arguments of the comparison. */
887 c->g1 = c->g2 = false;
891 c->u.s32.a = tcg_temp_new_i32();
892 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_dst);
893 c->u.s32.b = tcg_const_i32(0);
896 case CC_OP_LTUGTU_32:
898 c->u.s32.a = tcg_temp_new_i32();
899 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_src);
900 c->u.s32.b = tcg_temp_new_i32();
901 tcg_gen_trunc_i64_i32(c->u.s32.b, cc_dst);
908 c->u.s64.b = tcg_const_i64(0);
912 case CC_OP_LTUGTU_64:
915 c->g1 = c->g2 = true;
921 c->u.s64.a = tcg_temp_new_i64();
922 c->u.s64.b = tcg_const_i64(0);
923 tcg_gen_and_i64(c->u.s64.a, cc_src, cc_dst);
931 case 0x8 | 0x4 | 0x2: /* cc != 3 */
933 c->u.s32.b = tcg_const_i32(3);
935 case 0x8 | 0x4 | 0x1: /* cc != 2 */
937 c->u.s32.b = tcg_const_i32(2);
939 case 0x8 | 0x2 | 0x1: /* cc != 1 */
941 c->u.s32.b = tcg_const_i32(1);
943 case 0x8 | 0x2: /* cc == 0 ||Â cc == 2 => (cc & 1) == 0 */
946 c->u.s32.a = tcg_temp_new_i32();
947 c->u.s32.b = tcg_const_i32(0);
948 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
950 case 0x8 | 0x4: /* cc < 2 */
952 c->u.s32.b = tcg_const_i32(2);
954 case 0x8: /* cc == 0 */
956 c->u.s32.b = tcg_const_i32(0);
958 case 0x4 | 0x2 | 0x1: /* cc != 0 */
960 c->u.s32.b = tcg_const_i32(0);
962 case 0x4 | 0x1: /* cc == 1 ||Â cc == 3 => (cc & 1) != 0 */
965 c->u.s32.a = tcg_temp_new_i32();
966 c->u.s32.b = tcg_const_i32(0);
967 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
969 case 0x4: /* cc == 1 */
971 c->u.s32.b = tcg_const_i32(1);
973 case 0x2 | 0x1: /* cc > 1 */
975 c->u.s32.b = tcg_const_i32(1);
977 case 0x2: /* cc == 2 */
979 c->u.s32.b = tcg_const_i32(2);
981 case 0x1: /* cc == 3 */
983 c->u.s32.b = tcg_const_i32(3);
986 /* CC is masked by something else: (8 >> cc) & mask. */
989 c->u.s32.a = tcg_const_i32(8);
990 c->u.s32.b = tcg_const_i32(0);
991 tcg_gen_shr_i32(c->u.s32.a, c->u.s32.a, cc_op);
992 tcg_gen_andi_i32(c->u.s32.a, c->u.s32.a, mask);
1003 static void free_compare(DisasCompare *c)
1007 tcg_temp_free_i64(c->u.s64.a);
1009 tcg_temp_free_i32(c->u.s32.a);
1014 tcg_temp_free_i64(c->u.s64.b);
1016 tcg_temp_free_i32(c->u.s32.b);
1021 static void disas_b2(CPUS390XState *env, DisasContext *s, int op,
1024 #ifndef CONFIG_USER_ONLY
1029 r1 = (insn >> 4) & 0xf;
1032 LOG_DISAS("disas_b2: op 0x%x r1 %d r2 %d\n", op, r1, r2);
1035 case 0x20: /* SERVC R1,R2 [RRE] */
1036 /* SCLP Service call (PV hypercall) */
1037 check_privileged(s);
1038 potential_page_fault(s);
1039 tmp32_1 = load_reg32(r2);
1041 gen_helper_servc(cc_op, cpu_env, tmp32_1, tmp);
1043 tcg_temp_free_i32(tmp32_1);
1044 tcg_temp_free_i64(tmp);
1048 LOG_DISAS("illegal b2 operation 0x%x\n", op);
1049 gen_illegal_opcode(s);
1050 #ifndef CONFIG_USER_ONLY
1056 static void disas_s390_insn(CPUS390XState *env, DisasContext *s)
1062 opc = cpu_ldub_code(env, s->pc);
1063 LOG_DISAS("opc 0x%x\n", opc);
1067 insn = ld_code4(env, s->pc);
1068 op = (insn >> 16) & 0xff;
1069 disas_b2(env, s, op, insn);
1072 qemu_log_mask(LOG_UNIMP, "unimplemented opcode 0x%x\n", opc);
1073 gen_illegal_opcode(s);
1078 /* ====================================================================== */
1079 /* Define the insn format enumeration. */
1080 #define F0(N) FMT_##N,
1081 #define F1(N, X1) F0(N)
1082 #define F2(N, X1, X2) F0(N)
1083 #define F3(N, X1, X2, X3) F0(N)
1084 #define F4(N, X1, X2, X3, X4) F0(N)
1085 #define F5(N, X1, X2, X3, X4, X5) F0(N)
1088 #include "insn-format.def"
1098 /* Define a structure to hold the decoded fields. We'll store each inside
1099 an array indexed by an enum. In order to conserve memory, we'll arrange
1100 for fields that do not exist at the same time to overlap, thus the "C"
1101 for compact. For checking purposes there is an "O" for original index
1102 as well that will be applied to availability bitmaps. */
1104 enum DisasFieldIndexO {
1127 enum DisasFieldIndexC {
1158 struct DisasFields {
1161 unsigned presentC:16;
1162 unsigned int presentO;
1166 /* This is the way fields are to be accessed out of DisasFields. */
1167 #define have_field(S, F) have_field1((S), FLD_O_##F)
1168 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
1170 static bool have_field1(const DisasFields *f, enum DisasFieldIndexO c)
1172 return (f->presentO >> c) & 1;
1175 static int get_field1(const DisasFields *f, enum DisasFieldIndexO o,
1176 enum DisasFieldIndexC c)
1178 assert(have_field1(f, o));
1182 /* Describe the layout of each field in each format. */
1183 typedef struct DisasField {
1185 unsigned int size:8;
1186 unsigned int type:2;
1187 unsigned int indexC:6;
1188 enum DisasFieldIndexO indexO:8;
1191 typedef struct DisasFormatInfo {
1192 DisasField op[NUM_C_FIELD];
1195 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
1196 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
1197 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1198 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
1199 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1200 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1201 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
1202 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1203 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1204 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1205 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1206 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1207 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
1208 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
1210 #define F0(N) { { } },
1211 #define F1(N, X1) { { X1 } },
1212 #define F2(N, X1, X2) { { X1, X2 } },
1213 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
1214 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
1215 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
1217 static const DisasFormatInfo format_info[] = {
1218 #include "insn-format.def"
1236 /* Generally, we'll extract operands into this structures, operate upon
1237 them, and store them back. See the "in1", "in2", "prep", "wout" sets
1238 of routines below for more details. */
1240 bool g_out, g_out2, g_in1, g_in2;
1241 TCGv_i64 out, out2, in1, in2;
1245 /* Return values from translate_one, indicating the state of the TB. */
1247 /* Continue the TB. */
1249 /* We have emitted one or more goto_tb. No fixup required. */
1251 /* We are not using a goto_tb (for whatever reason), but have updated
1252 the PC (for whatever reason), so there's no need to do it again on
1255 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1256 updated the PC for the next instruction to be executed. */
1258 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1259 No following code will be executed. */
1263 typedef enum DisasFacility {
1264 FAC_Z, /* zarch (default) */
1265 FAC_CASS, /* compare and swap and store */
1266 FAC_CASS2, /* compare and swap and store 2*/
1267 FAC_DFP, /* decimal floating point */
1268 FAC_DFPR, /* decimal floating point rounding */
1269 FAC_DO, /* distinct operands */
1270 FAC_EE, /* execute extensions */
1271 FAC_EI, /* extended immediate */
1272 FAC_FPE, /* floating point extension */
1273 FAC_FPSSH, /* floating point support sign handling */
1274 FAC_FPRGR, /* FPR-GR transfer */
1275 FAC_GIE, /* general instructions extension */
1276 FAC_HFP_MA, /* HFP multiply-and-add/subtract */
1277 FAC_HW, /* high-word */
1278 FAC_IEEEE_SIM, /* IEEE exception sumilation */
1279 FAC_LOC, /* load/store on condition */
1280 FAC_LD, /* long displacement */
1281 FAC_PC, /* population count */
1282 FAC_SCF, /* store clock fast */
1283 FAC_SFLE, /* store facility list extended */
1289 DisasFacility fac:6;
1293 void (*help_in1)(DisasContext *, DisasFields *, DisasOps *);
1294 void (*help_in2)(DisasContext *, DisasFields *, DisasOps *);
1295 void (*help_prep)(DisasContext *, DisasFields *, DisasOps *);
1296 void (*help_wout)(DisasContext *, DisasFields *, DisasOps *);
1297 void (*help_cout)(DisasContext *, DisasOps *);
1298 ExitStatus (*help_op)(DisasContext *, DisasOps *);
1303 /* ====================================================================== */
1304 /* Miscelaneous helpers, used by several operations. */
1306 static void help_l2_shift(DisasContext *s, DisasFields *f,
1307 DisasOps *o, int mask)
1309 int b2 = get_field(f, b2);
1310 int d2 = get_field(f, d2);
1313 o->in2 = tcg_const_i64(d2 & mask);
1315 o->in2 = get_address(s, 0, b2, d2);
1316 tcg_gen_andi_i64(o->in2, o->in2, mask);
1320 static ExitStatus help_goto_direct(DisasContext *s, uint64_t dest)
1322 if (dest == s->next_pc) {
1325 if (use_goto_tb(s, dest)) {
1326 gen_update_cc_op(s);
1328 tcg_gen_movi_i64(psw_addr, dest);
1329 tcg_gen_exit_tb((tcg_target_long)s->tb);
1330 return EXIT_GOTO_TB;
1332 tcg_gen_movi_i64(psw_addr, dest);
1333 return EXIT_PC_UPDATED;
1337 static ExitStatus help_branch(DisasContext *s, DisasCompare *c,
1338 bool is_imm, int imm, TCGv_i64 cdest)
1341 uint64_t dest = s->pc + 2 * imm;
1344 /* Take care of the special cases first. */
1345 if (c->cond == TCG_COND_NEVER) {
1350 if (dest == s->next_pc) {
1351 /* Branch to next. */
1355 if (c->cond == TCG_COND_ALWAYS) {
1356 ret = help_goto_direct(s, dest);
1360 if (TCGV_IS_UNUSED_I64(cdest)) {
1361 /* E.g. bcr %r0 -> no branch. */
1365 if (c->cond == TCG_COND_ALWAYS) {
1366 tcg_gen_mov_i64(psw_addr, cdest);
1367 ret = EXIT_PC_UPDATED;
1372 if (use_goto_tb(s, s->next_pc)) {
1373 if (is_imm && use_goto_tb(s, dest)) {
1374 /* Both exits can use goto_tb. */
1375 gen_update_cc_op(s);
1377 lab = gen_new_label();
1379 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1381 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1384 /* Branch not taken. */
1386 tcg_gen_movi_i64(psw_addr, s->next_pc);
1387 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1392 tcg_gen_movi_i64(psw_addr, dest);
1393 tcg_gen_exit_tb((tcg_target_long)s->tb + 1);
1397 /* Fallthru can use goto_tb, but taken branch cannot. */
1398 /* Store taken branch destination before the brcond. This
1399 avoids having to allocate a new local temp to hold it.
1400 We'll overwrite this in the not taken case anyway. */
1402 tcg_gen_mov_i64(psw_addr, cdest);
1405 lab = gen_new_label();
1407 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1409 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1412 /* Branch not taken. */
1413 gen_update_cc_op(s);
1415 tcg_gen_movi_i64(psw_addr, s->next_pc);
1416 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1420 tcg_gen_movi_i64(psw_addr, dest);
1422 ret = EXIT_PC_UPDATED;
1425 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1426 Most commonly we're single-stepping or some other condition that
1427 disables all use of goto_tb. Just update the PC and exit. */
1429 TCGv_i64 next = tcg_const_i64(s->next_pc);
1431 cdest = tcg_const_i64(dest);
1435 tcg_gen_movcond_i64(c->cond, psw_addr, c->u.s64.a, c->u.s64.b,
1438 TCGv_i32 t0 = tcg_temp_new_i32();
1439 TCGv_i64 t1 = tcg_temp_new_i64();
1440 TCGv_i64 z = tcg_const_i64(0);
1441 tcg_gen_setcond_i32(c->cond, t0, c->u.s32.a, c->u.s32.b);
1442 tcg_gen_extu_i32_i64(t1, t0);
1443 tcg_temp_free_i32(t0);
1444 tcg_gen_movcond_i64(TCG_COND_NE, psw_addr, t1, z, cdest, next);
1445 tcg_temp_free_i64(t1);
1446 tcg_temp_free_i64(z);
1450 tcg_temp_free_i64(cdest);
1452 tcg_temp_free_i64(next);
1454 ret = EXIT_PC_UPDATED;
1462 /* ====================================================================== */
1463 /* The operations. These perform the bulk of the work for any insn,
1464 usually after the operands have been loaded and output initialized. */
1466 static ExitStatus op_abs(DisasContext *s, DisasOps *o)
1468 gen_helper_abs_i64(o->out, o->in2);
1472 static ExitStatus op_absf32(DisasContext *s, DisasOps *o)
1474 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffull);
1478 static ExitStatus op_absf64(DisasContext *s, DisasOps *o)
1480 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffffffffffull);
1484 static ExitStatus op_absf128(DisasContext *s, DisasOps *o)
1486 tcg_gen_andi_i64(o->out, o->in1, 0x7fffffffffffffffull);
1487 tcg_gen_mov_i64(o->out2, o->in2);
1491 static ExitStatus op_add(DisasContext *s, DisasOps *o)
1493 tcg_gen_add_i64(o->out, o->in1, o->in2);
1497 static ExitStatus op_addc(DisasContext *s, DisasOps *o)
1501 tcg_gen_add_i64(o->out, o->in1, o->in2);
1503 /* XXX possible optimization point */
1505 cc = tcg_temp_new_i64();
1506 tcg_gen_extu_i32_i64(cc, cc_op);
1507 tcg_gen_shri_i64(cc, cc, 1);
1509 tcg_gen_add_i64(o->out, o->out, cc);
1510 tcg_temp_free_i64(cc);
1514 static ExitStatus op_aeb(DisasContext *s, DisasOps *o)
1516 gen_helper_aeb(o->out, cpu_env, o->in1, o->in2);
1520 static ExitStatus op_adb(DisasContext *s, DisasOps *o)
1522 gen_helper_adb(o->out, cpu_env, o->in1, o->in2);
1526 static ExitStatus op_axb(DisasContext *s, DisasOps *o)
1528 gen_helper_axb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1529 return_low128(o->out2);
1533 static ExitStatus op_and(DisasContext *s, DisasOps *o)
1535 tcg_gen_and_i64(o->out, o->in1, o->in2);
1539 static ExitStatus op_andi(DisasContext *s, DisasOps *o)
1541 int shift = s->insn->data & 0xff;
1542 int size = s->insn->data >> 8;
1543 uint64_t mask = ((1ull << size) - 1) << shift;
1546 tcg_gen_shli_i64(o->in2, o->in2, shift);
1547 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
1548 tcg_gen_and_i64(o->out, o->in1, o->in2);
1550 /* Produce the CC from only the bits manipulated. */
1551 tcg_gen_andi_i64(cc_dst, o->out, mask);
1552 set_cc_nz_u64(s, cc_dst);
1556 static ExitStatus op_bas(DisasContext *s, DisasOps *o)
1558 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1559 if (!TCGV_IS_UNUSED_I64(o->in2)) {
1560 tcg_gen_mov_i64(psw_addr, o->in2);
1561 return EXIT_PC_UPDATED;
1567 static ExitStatus op_basi(DisasContext *s, DisasOps *o)
1569 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1570 return help_goto_direct(s, s->pc + 2 * get_field(s->fields, i2));
1573 static ExitStatus op_bc(DisasContext *s, DisasOps *o)
1575 int m1 = get_field(s->fields, m1);
1576 bool is_imm = have_field(s->fields, i2);
1577 int imm = is_imm ? get_field(s->fields, i2) : 0;
1580 disas_jcc(s, &c, m1);
1581 return help_branch(s, &c, is_imm, imm, o->in2);
1584 static ExitStatus op_bct32(DisasContext *s, DisasOps *o)
1586 int r1 = get_field(s->fields, r1);
1587 bool is_imm = have_field(s->fields, i2);
1588 int imm = is_imm ? get_field(s->fields, i2) : 0;
1592 c.cond = TCG_COND_NE;
1597 t = tcg_temp_new_i64();
1598 tcg_gen_subi_i64(t, regs[r1], 1);
1599 store_reg32_i64(r1, t);
1600 c.u.s32.a = tcg_temp_new_i32();
1601 c.u.s32.b = tcg_const_i32(0);
1602 tcg_gen_trunc_i64_i32(c.u.s32.a, t);
1603 tcg_temp_free_i64(t);
1605 return help_branch(s, &c, is_imm, imm, o->in2);
1608 static ExitStatus op_bct64(DisasContext *s, DisasOps *o)
1610 int r1 = get_field(s->fields, r1);
1611 bool is_imm = have_field(s->fields, i2);
1612 int imm = is_imm ? get_field(s->fields, i2) : 0;
1615 c.cond = TCG_COND_NE;
1620 tcg_gen_subi_i64(regs[r1], regs[r1], 1);
1621 c.u.s64.a = regs[r1];
1622 c.u.s64.b = tcg_const_i64(0);
1624 return help_branch(s, &c, is_imm, imm, o->in2);
1627 static ExitStatus op_ceb(DisasContext *s, DisasOps *o)
1629 gen_helper_ceb(cc_op, cpu_env, o->in1, o->in2);
1634 static ExitStatus op_cdb(DisasContext *s, DisasOps *o)
1636 gen_helper_cdb(cc_op, cpu_env, o->in1, o->in2);
1641 static ExitStatus op_cxb(DisasContext *s, DisasOps *o)
1643 gen_helper_cxb(cc_op, cpu_env, o->out, o->out2, o->in1, o->in2);
1648 static ExitStatus op_cfeb(DisasContext *s, DisasOps *o)
1650 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1651 gen_helper_cfeb(o->out, cpu_env, o->in2, m3);
1652 tcg_temp_free_i32(m3);
1653 gen_set_cc_nz_f32(s, o->in2);
1657 static ExitStatus op_cfdb(DisasContext *s, DisasOps *o)
1659 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1660 gen_helper_cfdb(o->out, cpu_env, o->in2, m3);
1661 tcg_temp_free_i32(m3);
1662 gen_set_cc_nz_f64(s, o->in2);
1666 static ExitStatus op_cfxb(DisasContext *s, DisasOps *o)
1668 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1669 gen_helper_cfxb(o->out, cpu_env, o->in1, o->in2, m3);
1670 tcg_temp_free_i32(m3);
1671 gen_set_cc_nz_f128(s, o->in1, o->in2);
1675 static ExitStatus op_cgeb(DisasContext *s, DisasOps *o)
1677 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1678 gen_helper_cgeb(o->out, cpu_env, o->in2, m3);
1679 tcg_temp_free_i32(m3);
1680 gen_set_cc_nz_f32(s, o->in2);
1684 static ExitStatus op_cgdb(DisasContext *s, DisasOps *o)
1686 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1687 gen_helper_cgdb(o->out, cpu_env, o->in2, m3);
1688 tcg_temp_free_i32(m3);
1689 gen_set_cc_nz_f64(s, o->in2);
1693 static ExitStatus op_cgxb(DisasContext *s, DisasOps *o)
1695 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1696 gen_helper_cgxb(o->out, cpu_env, o->in1, o->in2, m3);
1697 tcg_temp_free_i32(m3);
1698 gen_set_cc_nz_f128(s, o->in1, o->in2);
1702 static ExitStatus op_cegb(DisasContext *s, DisasOps *o)
1704 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1705 gen_helper_cegb(o->out, cpu_env, o->in2, m3);
1706 tcg_temp_free_i32(m3);
1710 static ExitStatus op_cdgb(DisasContext *s, DisasOps *o)
1712 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1713 gen_helper_cdgb(o->out, cpu_env, o->in2, m3);
1714 tcg_temp_free_i32(m3);
1718 static ExitStatus op_cxgb(DisasContext *s, DisasOps *o)
1720 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1721 gen_helper_cxgb(o->out, cpu_env, o->in2, m3);
1722 tcg_temp_free_i32(m3);
1723 return_low128(o->out2);
1727 static ExitStatus op_cksm(DisasContext *s, DisasOps *o)
1729 int r2 = get_field(s->fields, r2);
1730 TCGv_i64 len = tcg_temp_new_i64();
1732 potential_page_fault(s);
1733 gen_helper_cksm(len, cpu_env, o->in1, o->in2, regs[r2 + 1]);
1735 return_low128(o->out);
1737 tcg_gen_add_i64(regs[r2], regs[r2], len);
1738 tcg_gen_sub_i64(regs[r2 + 1], regs[r2 + 1], len);
1739 tcg_temp_free_i64(len);
1744 static ExitStatus op_clc(DisasContext *s, DisasOps *o)
1746 int l = get_field(s->fields, l1);
1751 tcg_gen_qemu_ld8u(cc_src, o->addr1, get_mem_index(s));
1752 tcg_gen_qemu_ld8u(cc_dst, o->in2, get_mem_index(s));
1755 tcg_gen_qemu_ld16u(cc_src, o->addr1, get_mem_index(s));
1756 tcg_gen_qemu_ld16u(cc_dst, o->in2, get_mem_index(s));
1759 tcg_gen_qemu_ld32u(cc_src, o->addr1, get_mem_index(s));
1760 tcg_gen_qemu_ld32u(cc_dst, o->in2, get_mem_index(s));
1763 tcg_gen_qemu_ld64(cc_src, o->addr1, get_mem_index(s));
1764 tcg_gen_qemu_ld64(cc_dst, o->in2, get_mem_index(s));
1767 potential_page_fault(s);
1768 vl = tcg_const_i32(l);
1769 gen_helper_clc(cc_op, cpu_env, vl, o->addr1, o->in2);
1770 tcg_temp_free_i32(vl);
1774 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, cc_src, cc_dst);
1778 static ExitStatus op_clcle(DisasContext *s, DisasOps *o)
1780 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1781 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1782 potential_page_fault(s);
1783 gen_helper_clcle(cc_op, cpu_env, r1, o->in2, r3);
1784 tcg_temp_free_i32(r1);
1785 tcg_temp_free_i32(r3);
1790 static ExitStatus op_clm(DisasContext *s, DisasOps *o)
1792 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1793 TCGv_i32 t1 = tcg_temp_new_i32();
1794 tcg_gen_trunc_i64_i32(t1, o->in1);
1795 potential_page_fault(s);
1796 gen_helper_clm(cc_op, cpu_env, t1, m3, o->in2);
1798 tcg_temp_free_i32(t1);
1799 tcg_temp_free_i32(m3);
1803 static ExitStatus op_clst(DisasContext *s, DisasOps *o)
1805 potential_page_fault(s);
1806 gen_helper_clst(o->in1, cpu_env, regs[0], o->in1, o->in2);
1808 return_low128(o->in2);
1812 static ExitStatus op_cs(DisasContext *s, DisasOps *o)
1814 int r3 = get_field(s->fields, r3);
1815 potential_page_fault(s);
1816 gen_helper_cs(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1821 static ExitStatus op_csg(DisasContext *s, DisasOps *o)
1823 int r3 = get_field(s->fields, r3);
1824 potential_page_fault(s);
1825 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1830 #ifndef CONFIG_USER_ONLY
1831 static ExitStatus op_csp(DisasContext *s, DisasOps *o)
1833 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1834 check_privileged(s);
1835 gen_helper_csp(cc_op, cpu_env, r1, o->in2);
1836 tcg_temp_free_i32(r1);
1842 static ExitStatus op_cds(DisasContext *s, DisasOps *o)
1844 int r3 = get_field(s->fields, r3);
1845 TCGv_i64 in3 = tcg_temp_new_i64();
1846 tcg_gen_deposit_i64(in3, regs[r3 + 1], regs[r3], 32, 32);
1847 potential_page_fault(s);
1848 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, in3);
1849 tcg_temp_free_i64(in3);
1854 static ExitStatus op_cdsg(DisasContext *s, DisasOps *o)
1856 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1857 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1858 potential_page_fault(s);
1859 /* XXX rewrite in tcg */
1860 gen_helper_cdsg(cc_op, cpu_env, r1, o->in2, r3);
1865 static ExitStatus op_cvd(DisasContext *s, DisasOps *o)
1867 TCGv_i64 t1 = tcg_temp_new_i64();
1868 TCGv_i32 t2 = tcg_temp_new_i32();
1869 tcg_gen_trunc_i64_i32(t2, o->in1);
1870 gen_helper_cvd(t1, t2);
1871 tcg_temp_free_i32(t2);
1872 tcg_gen_qemu_st64(t1, o->in2, get_mem_index(s));
1873 tcg_temp_free_i64(t1);
1877 #ifndef CONFIG_USER_ONLY
1878 static ExitStatus op_diag(DisasContext *s, DisasOps *o)
1882 check_privileged(s);
1883 potential_page_fault(s);
1885 /* We pretend the format is RX_a so that D2 is the field we want. */
1886 tmp = tcg_const_i32(get_field(s->fields, d2) & 0xfff);
1887 gen_helper_diag(regs[2], cpu_env, tmp, regs[2], regs[1]);
1888 tcg_temp_free_i32(tmp);
1893 static ExitStatus op_divs32(DisasContext *s, DisasOps *o)
1895 gen_helper_divs32(o->out2, cpu_env, o->in1, o->in2);
1896 return_low128(o->out);
1900 static ExitStatus op_divu32(DisasContext *s, DisasOps *o)
1902 gen_helper_divu32(o->out2, cpu_env, o->in1, o->in2);
1903 return_low128(o->out);
1907 static ExitStatus op_divs64(DisasContext *s, DisasOps *o)
1909 gen_helper_divs64(o->out2, cpu_env, o->in1, o->in2);
1910 return_low128(o->out);
1914 static ExitStatus op_divu64(DisasContext *s, DisasOps *o)
1916 gen_helper_divu64(o->out2, cpu_env, o->out, o->out2, o->in2);
1917 return_low128(o->out);
1921 static ExitStatus op_deb(DisasContext *s, DisasOps *o)
1923 gen_helper_deb(o->out, cpu_env, o->in1, o->in2);
1927 static ExitStatus op_ddb(DisasContext *s, DisasOps *o)
1929 gen_helper_ddb(o->out, cpu_env, o->in1, o->in2);
1933 static ExitStatus op_dxb(DisasContext *s, DisasOps *o)
1935 gen_helper_dxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1936 return_low128(o->out2);
1940 static ExitStatus op_ear(DisasContext *s, DisasOps *o)
1942 int r2 = get_field(s->fields, r2);
1943 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, aregs[r2]));
1947 static ExitStatus op_efpc(DisasContext *s, DisasOps *o)
1949 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, fpc));
1953 static ExitStatus op_ex(DisasContext *s, DisasOps *o)
1955 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
1956 tb->flags, (ab)use the tb->cs_base field as the address of
1957 the template in memory, and grab 8 bits of tb->flags/cflags for
1958 the contents of the register. We would then recognize all this
1959 in gen_intermediate_code_internal, generating code for exactly
1960 one instruction. This new TB then gets executed normally.
1962 On the other hand, this seems to be mostly used for modifying
1963 MVC inside of memcpy, which needs a helper call anyway. So
1964 perhaps this doesn't bear thinking about any further. */
1971 tmp = tcg_const_i64(s->next_pc);
1972 gen_helper_ex(cc_op, cpu_env, cc_op, o->in1, o->in2, tmp);
1973 tcg_temp_free_i64(tmp);
1979 static ExitStatus op_flogr(DisasContext *s, DisasOps *o)
1981 /* We'll use the original input for cc computation, since we get to
1982 compare that against 0, which ought to be better than comparing
1983 the real output against 64. It also lets cc_dst be a convenient
1984 temporary during our computation. */
1985 gen_op_update1_cc_i64(s, CC_OP_FLOGR, o->in2);
1987 /* R1 = IN ? CLZ(IN) : 64. */
1988 gen_helper_clz(o->out, o->in2);
1990 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
1991 value by 64, which is undefined. But since the shift is 64 iff the
1992 input is zero, we still get the correct result after and'ing. */
1993 tcg_gen_movi_i64(o->out2, 0x8000000000000000ull);
1994 tcg_gen_shr_i64(o->out2, o->out2, o->out);
1995 tcg_gen_andc_i64(o->out2, cc_dst, o->out2);
1999 static ExitStatus op_icm(DisasContext *s, DisasOps *o)
2001 int m3 = get_field(s->fields, m3);
2002 int pos, len, base = s->insn->data;
2003 TCGv_i64 tmp = tcg_temp_new_i64();
2008 /* Effectively a 32-bit load. */
2009 tcg_gen_qemu_ld32u(tmp, o->in2, get_mem_index(s));
2016 /* Effectively a 16-bit load. */
2017 tcg_gen_qemu_ld16u(tmp, o->in2, get_mem_index(s));
2025 /* Effectively an 8-bit load. */
2026 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2031 pos = base + ctz32(m3) * 8;
2032 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, len);
2033 ccm = ((1ull << len) - 1) << pos;
2037 /* This is going to be a sequence of loads and inserts. */
2038 pos = base + 32 - 8;
2042 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2043 tcg_gen_addi_i64(o->in2, o->in2, 1);
2044 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, 8);
2047 m3 = (m3 << 1) & 0xf;
2053 tcg_gen_movi_i64(tmp, ccm);
2054 gen_op_update2_cc_i64(s, CC_OP_ICM, tmp, o->out);
2055 tcg_temp_free_i64(tmp);
2059 static ExitStatus op_insi(DisasContext *s, DisasOps *o)
2061 int shift = s->insn->data & 0xff;
2062 int size = s->insn->data >> 8;
2063 tcg_gen_deposit_i64(o->out, o->in1, o->in2, shift, size);
2067 static ExitStatus op_ipm(DisasContext *s, DisasOps *o)
2072 tcg_gen_andi_i64(o->out, o->out, ~0xff000000ull);
2074 t1 = tcg_temp_new_i64();
2075 tcg_gen_shli_i64(t1, psw_mask, 20);
2076 tcg_gen_shri_i64(t1, t1, 36);
2077 tcg_gen_or_i64(o->out, o->out, t1);
2079 tcg_gen_extu_i32_i64(t1, cc_op);
2080 tcg_gen_shli_i64(t1, t1, 28);
2081 tcg_gen_or_i64(o->out, o->out, t1);
2082 tcg_temp_free_i64(t1);
2086 #ifndef CONFIG_USER_ONLY
2087 static ExitStatus op_ipte(DisasContext *s, DisasOps *o)
2089 check_privileged(s);
2090 gen_helper_ipte(cpu_env, o->in1, o->in2);
2094 static ExitStatus op_iske(DisasContext *s, DisasOps *o)
2096 check_privileged(s);
2097 gen_helper_iske(o->out, cpu_env, o->in2);
2102 static ExitStatus op_ldeb(DisasContext *s, DisasOps *o)
2104 gen_helper_ldeb(o->out, cpu_env, o->in2);
2108 static ExitStatus op_ledb(DisasContext *s, DisasOps *o)
2110 gen_helper_ledb(o->out, cpu_env, o->in2);
2114 static ExitStatus op_ldxb(DisasContext *s, DisasOps *o)
2116 gen_helper_ldxb(o->out, cpu_env, o->in1, o->in2);
2120 static ExitStatus op_lexb(DisasContext *s, DisasOps *o)
2122 gen_helper_lexb(o->out, cpu_env, o->in1, o->in2);
2126 static ExitStatus op_lxdb(DisasContext *s, DisasOps *o)
2128 gen_helper_lxdb(o->out, cpu_env, o->in2);
2129 return_low128(o->out2);
2133 static ExitStatus op_lxeb(DisasContext *s, DisasOps *o)
2135 gen_helper_lxeb(o->out, cpu_env, o->in2);
2136 return_low128(o->out2);
2140 static ExitStatus op_llgt(DisasContext *s, DisasOps *o)
2142 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffff);
2146 static ExitStatus op_ld8s(DisasContext *s, DisasOps *o)
2148 tcg_gen_qemu_ld8s(o->out, o->in2, get_mem_index(s));
2152 static ExitStatus op_ld8u(DisasContext *s, DisasOps *o)
2154 tcg_gen_qemu_ld8u(o->out, o->in2, get_mem_index(s));
2158 static ExitStatus op_ld16s(DisasContext *s, DisasOps *o)
2160 tcg_gen_qemu_ld16s(o->out, o->in2, get_mem_index(s));
2164 static ExitStatus op_ld16u(DisasContext *s, DisasOps *o)
2166 tcg_gen_qemu_ld16u(o->out, o->in2, get_mem_index(s));
2170 static ExitStatus op_ld32s(DisasContext *s, DisasOps *o)
2172 tcg_gen_qemu_ld32s(o->out, o->in2, get_mem_index(s));
2176 static ExitStatus op_ld32u(DisasContext *s, DisasOps *o)
2178 tcg_gen_qemu_ld32u(o->out, o->in2, get_mem_index(s));
2182 static ExitStatus op_ld64(DisasContext *s, DisasOps *o)
2184 tcg_gen_qemu_ld64(o->out, o->in2, get_mem_index(s));
2188 #ifndef CONFIG_USER_ONLY
2189 static ExitStatus op_lctl(DisasContext *s, DisasOps *o)
2191 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2192 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2193 check_privileged(s);
2194 potential_page_fault(s);
2195 gen_helper_lctl(cpu_env, r1, o->in2, r3);
2196 tcg_temp_free_i32(r1);
2197 tcg_temp_free_i32(r3);
2201 static ExitStatus op_lctlg(DisasContext *s, DisasOps *o)
2203 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2204 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2205 check_privileged(s);
2206 potential_page_fault(s);
2207 gen_helper_lctlg(cpu_env, r1, o->in2, r3);
2208 tcg_temp_free_i32(r1);
2209 tcg_temp_free_i32(r3);
2212 static ExitStatus op_lra(DisasContext *s, DisasOps *o)
2214 check_privileged(s);
2215 potential_page_fault(s);
2216 gen_helper_lra(o->out, cpu_env, o->in2);
2221 static ExitStatus op_lpsw(DisasContext *s, DisasOps *o)
2225 check_privileged(s);
2227 t1 = tcg_temp_new_i64();
2228 t2 = tcg_temp_new_i64();
2229 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2230 tcg_gen_addi_i64(o->in2, o->in2, 4);
2231 tcg_gen_qemu_ld32u(t2, o->in2, get_mem_index(s));
2232 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2233 tcg_gen_shli_i64(t1, t1, 32);
2234 gen_helper_load_psw(cpu_env, t1, t2);
2235 tcg_temp_free_i64(t1);
2236 tcg_temp_free_i64(t2);
2237 return EXIT_NORETURN;
2240 static ExitStatus op_lpswe(DisasContext *s, DisasOps *o)
2244 check_privileged(s);
2246 t1 = tcg_temp_new_i64();
2247 t2 = tcg_temp_new_i64();
2248 tcg_gen_qemu_ld64(t1, o->in2, get_mem_index(s));
2249 tcg_gen_addi_i64(o->in2, o->in2, 8);
2250 tcg_gen_qemu_ld64(t2, o->in2, get_mem_index(s));
2251 gen_helper_load_psw(cpu_env, t1, t2);
2252 tcg_temp_free_i64(t1);
2253 tcg_temp_free_i64(t2);
2254 return EXIT_NORETURN;
2258 static ExitStatus op_lam(DisasContext *s, DisasOps *o)
2260 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2261 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2262 potential_page_fault(s);
2263 gen_helper_lam(cpu_env, r1, o->in2, r3);
2264 tcg_temp_free_i32(r1);
2265 tcg_temp_free_i32(r3);
2269 static ExitStatus op_lm32(DisasContext *s, DisasOps *o)
2271 int r1 = get_field(s->fields, r1);
2272 int r3 = get_field(s->fields, r3);
2273 TCGv_i64 t = tcg_temp_new_i64();
2274 TCGv_i64 t4 = tcg_const_i64(4);
2277 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2278 store_reg32_i64(r1, t);
2282 tcg_gen_add_i64(o->in2, o->in2, t4);
2286 tcg_temp_free_i64(t);
2287 tcg_temp_free_i64(t4);
2291 static ExitStatus op_lmh(DisasContext *s, DisasOps *o)
2293 int r1 = get_field(s->fields, r1);
2294 int r3 = get_field(s->fields, r3);
2295 TCGv_i64 t = tcg_temp_new_i64();
2296 TCGv_i64 t4 = tcg_const_i64(4);
2299 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2300 store_reg32h_i64(r1, t);
2304 tcg_gen_add_i64(o->in2, o->in2, t4);
2308 tcg_temp_free_i64(t);
2309 tcg_temp_free_i64(t4);
2313 static ExitStatus op_lm64(DisasContext *s, DisasOps *o)
2315 int r1 = get_field(s->fields, r1);
2316 int r3 = get_field(s->fields, r3);
2317 TCGv_i64 t8 = tcg_const_i64(8);
2320 tcg_gen_qemu_ld64(regs[r1], o->in2, get_mem_index(s));
2324 tcg_gen_add_i64(o->in2, o->in2, t8);
2328 tcg_temp_free_i64(t8);
2332 static ExitStatus op_mov2(DisasContext *s, DisasOps *o)
2335 o->g_out = o->g_in2;
2336 TCGV_UNUSED_I64(o->in2);
2341 static ExitStatus op_movx(DisasContext *s, DisasOps *o)
2345 o->g_out = o->g_in1;
2346 o->g_out2 = o->g_in2;
2347 TCGV_UNUSED_I64(o->in1);
2348 TCGV_UNUSED_I64(o->in2);
2349 o->g_in1 = o->g_in2 = false;
2353 static ExitStatus op_mvc(DisasContext *s, DisasOps *o)
2355 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2356 potential_page_fault(s);
2357 gen_helper_mvc(cpu_env, l, o->addr1, o->in2);
2358 tcg_temp_free_i32(l);
2362 static ExitStatus op_mvcl(DisasContext *s, DisasOps *o)
2364 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2365 TCGv_i32 r2 = tcg_const_i32(get_field(s->fields, r2));
2366 potential_page_fault(s);
2367 gen_helper_mvcl(cc_op, cpu_env, r1, r2);
2368 tcg_temp_free_i32(r1);
2369 tcg_temp_free_i32(r2);
2374 static ExitStatus op_mvcle(DisasContext *s, DisasOps *o)
2376 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2377 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2378 potential_page_fault(s);
2379 gen_helper_mvcle(cc_op, cpu_env, r1, o->in2, r3);
2380 tcg_temp_free_i32(r1);
2381 tcg_temp_free_i32(r3);
2386 #ifndef CONFIG_USER_ONLY
2387 static ExitStatus op_mvcp(DisasContext *s, DisasOps *o)
2389 int r1 = get_field(s->fields, l1);
2390 check_privileged(s);
2391 potential_page_fault(s);
2392 gen_helper_mvcp(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2397 static ExitStatus op_mvcs(DisasContext *s, DisasOps *o)
2399 int r1 = get_field(s->fields, l1);
2400 check_privileged(s);
2401 potential_page_fault(s);
2402 gen_helper_mvcs(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2408 static ExitStatus op_mvpg(DisasContext *s, DisasOps *o)
2410 potential_page_fault(s);
2411 gen_helper_mvpg(cpu_env, regs[0], o->in1, o->in2);
2416 static ExitStatus op_mvst(DisasContext *s, DisasOps *o)
2418 potential_page_fault(s);
2419 gen_helper_mvst(o->in1, cpu_env, regs[0], o->in1, o->in2);
2421 return_low128(o->in2);
2425 static ExitStatus op_mul(DisasContext *s, DisasOps *o)
2427 tcg_gen_mul_i64(o->out, o->in1, o->in2);
2431 static ExitStatus op_mul128(DisasContext *s, DisasOps *o)
2433 gen_helper_mul128(o->out, cpu_env, o->in1, o->in2);
2434 return_low128(o->out2);
2438 static ExitStatus op_meeb(DisasContext *s, DisasOps *o)
2440 gen_helper_meeb(o->out, cpu_env, o->in1, o->in2);
2444 static ExitStatus op_mdeb(DisasContext *s, DisasOps *o)
2446 gen_helper_mdeb(o->out, cpu_env, o->in1, o->in2);
2450 static ExitStatus op_mdb(DisasContext *s, DisasOps *o)
2452 gen_helper_mdb(o->out, cpu_env, o->in1, o->in2);
2456 static ExitStatus op_mxb(DisasContext *s, DisasOps *o)
2458 gen_helper_mxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2459 return_low128(o->out2);
2463 static ExitStatus op_mxdb(DisasContext *s, DisasOps *o)
2465 gen_helper_mxdb(o->out, cpu_env, o->out, o->out2, o->in2);
2466 return_low128(o->out2);
2470 static ExitStatus op_maeb(DisasContext *s, DisasOps *o)
2472 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2473 gen_helper_maeb(o->out, cpu_env, o->in1, o->in2, r3);
2474 tcg_temp_free_i64(r3);
2478 static ExitStatus op_madb(DisasContext *s, DisasOps *o)
2480 int r3 = get_field(s->fields, r3);
2481 gen_helper_madb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2485 static ExitStatus op_mseb(DisasContext *s, DisasOps *o)
2487 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2488 gen_helper_mseb(o->out, cpu_env, o->in1, o->in2, r3);
2489 tcg_temp_free_i64(r3);
2493 static ExitStatus op_msdb(DisasContext *s, DisasOps *o)
2495 int r3 = get_field(s->fields, r3);
2496 gen_helper_msdb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2500 static ExitStatus op_nabs(DisasContext *s, DisasOps *o)
2502 gen_helper_nabs_i64(o->out, o->in2);
2506 static ExitStatus op_nabsf32(DisasContext *s, DisasOps *o)
2508 tcg_gen_ori_i64(o->out, o->in2, 0x80000000ull);
2512 static ExitStatus op_nabsf64(DisasContext *s, DisasOps *o)
2514 tcg_gen_ori_i64(o->out, o->in2, 0x8000000000000000ull);
2518 static ExitStatus op_nabsf128(DisasContext *s, DisasOps *o)
2520 tcg_gen_ori_i64(o->out, o->in1, 0x8000000000000000ull);
2521 tcg_gen_mov_i64(o->out2, o->in2);
2525 static ExitStatus op_nc(DisasContext *s, DisasOps *o)
2527 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2528 potential_page_fault(s);
2529 gen_helper_nc(cc_op, cpu_env, l, o->addr1, o->in2);
2530 tcg_temp_free_i32(l);
2535 static ExitStatus op_neg(DisasContext *s, DisasOps *o)
2537 tcg_gen_neg_i64(o->out, o->in2);
2541 static ExitStatus op_negf32(DisasContext *s, DisasOps *o)
2543 tcg_gen_xori_i64(o->out, o->in2, 0x80000000ull);
2547 static ExitStatus op_negf64(DisasContext *s, DisasOps *o)
2549 tcg_gen_xori_i64(o->out, o->in2, 0x8000000000000000ull);
2553 static ExitStatus op_negf128(DisasContext *s, DisasOps *o)
2555 tcg_gen_xori_i64(o->out, o->in1, 0x8000000000000000ull);
2556 tcg_gen_mov_i64(o->out2, o->in2);
2560 static ExitStatus op_oc(DisasContext *s, DisasOps *o)
2562 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2563 potential_page_fault(s);
2564 gen_helper_oc(cc_op, cpu_env, l, o->addr1, o->in2);
2565 tcg_temp_free_i32(l);
2570 static ExitStatus op_or(DisasContext *s, DisasOps *o)
2572 tcg_gen_or_i64(o->out, o->in1, o->in2);
2576 static ExitStatus op_ori(DisasContext *s, DisasOps *o)
2578 int shift = s->insn->data & 0xff;
2579 int size = s->insn->data >> 8;
2580 uint64_t mask = ((1ull << size) - 1) << shift;
2583 tcg_gen_shli_i64(o->in2, o->in2, shift);
2584 tcg_gen_or_i64(o->out, o->in1, o->in2);
2586 /* Produce the CC from only the bits manipulated. */
2587 tcg_gen_andi_i64(cc_dst, o->out, mask);
2588 set_cc_nz_u64(s, cc_dst);
2592 #ifndef CONFIG_USER_ONLY
2593 static ExitStatus op_ptlb(DisasContext *s, DisasOps *o)
2595 check_privileged(s);
2596 gen_helper_ptlb(cpu_env);
2601 static ExitStatus op_rev16(DisasContext *s, DisasOps *o)
2603 tcg_gen_bswap16_i64(o->out, o->in2);
2607 static ExitStatus op_rev32(DisasContext *s, DisasOps *o)
2609 tcg_gen_bswap32_i64(o->out, o->in2);
2613 static ExitStatus op_rev64(DisasContext *s, DisasOps *o)
2615 tcg_gen_bswap64_i64(o->out, o->in2);
2619 static ExitStatus op_rll32(DisasContext *s, DisasOps *o)
2621 TCGv_i32 t1 = tcg_temp_new_i32();
2622 TCGv_i32 t2 = tcg_temp_new_i32();
2623 TCGv_i32 to = tcg_temp_new_i32();
2624 tcg_gen_trunc_i64_i32(t1, o->in1);
2625 tcg_gen_trunc_i64_i32(t2, o->in2);
2626 tcg_gen_rotl_i32(to, t1, t2);
2627 tcg_gen_extu_i32_i64(o->out, to);
2628 tcg_temp_free_i32(t1);
2629 tcg_temp_free_i32(t2);
2630 tcg_temp_free_i32(to);
2634 static ExitStatus op_rll64(DisasContext *s, DisasOps *o)
2636 tcg_gen_rotl_i64(o->out, o->in1, o->in2);
2640 #ifndef CONFIG_USER_ONLY
2641 static ExitStatus op_rrbe(DisasContext *s, DisasOps *o)
2643 check_privileged(s);
2644 gen_helper_rrbe(cc_op, cpu_env, o->in2);
2649 static ExitStatus op_sacf(DisasContext *s, DisasOps *o)
2651 check_privileged(s);
2652 gen_helper_sacf(cpu_env, o->in2);
2653 /* Addressing mode has changed, so end the block. */
2654 return EXIT_PC_STALE;
2658 static ExitStatus op_sar(DisasContext *s, DisasOps *o)
2660 int r1 = get_field(s->fields, r1);
2661 tcg_gen_st32_i64(o->in2, cpu_env, offsetof(CPUS390XState, aregs[r1]));
2665 static ExitStatus op_seb(DisasContext *s, DisasOps *o)
2667 gen_helper_seb(o->out, cpu_env, o->in1, o->in2);
2671 static ExitStatus op_sdb(DisasContext *s, DisasOps *o)
2673 gen_helper_sdb(o->out, cpu_env, o->in1, o->in2);
2677 static ExitStatus op_sxb(DisasContext *s, DisasOps *o)
2679 gen_helper_sxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2680 return_low128(o->out2);
2684 static ExitStatus op_sqeb(DisasContext *s, DisasOps *o)
2686 gen_helper_sqeb(o->out, cpu_env, o->in2);
2690 static ExitStatus op_sqdb(DisasContext *s, DisasOps *o)
2692 gen_helper_sqdb(o->out, cpu_env, o->in2);
2696 static ExitStatus op_sqxb(DisasContext *s, DisasOps *o)
2698 gen_helper_sqxb(o->out, cpu_env, o->in1, o->in2);
2699 return_low128(o->out2);
2703 #ifndef CONFIG_USER_ONLY
2704 static ExitStatus op_sigp(DisasContext *s, DisasOps *o)
2706 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2707 check_privileged(s);
2708 potential_page_fault(s);
2709 gen_helper_sigp(cc_op, cpu_env, o->in2, r1, o->in1);
2710 tcg_temp_free_i32(r1);
2715 static ExitStatus op_sla(DisasContext *s, DisasOps *o)
2717 uint64_t sign = 1ull << s->insn->data;
2718 enum cc_op cco = s->insn->data == 31 ? CC_OP_SLA_32 : CC_OP_SLA_64;
2719 gen_op_update2_cc_i64(s, cco, o->in1, o->in2);
2720 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2721 /* The arithmetic left shift is curious in that it does not affect
2722 the sign bit. Copy that over from the source unchanged. */
2723 tcg_gen_andi_i64(o->out, o->out, ~sign);
2724 tcg_gen_andi_i64(o->in1, o->in1, sign);
2725 tcg_gen_or_i64(o->out, o->out, o->in1);
2729 static ExitStatus op_sll(DisasContext *s, DisasOps *o)
2731 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2735 static ExitStatus op_sra(DisasContext *s, DisasOps *o)
2737 tcg_gen_sar_i64(o->out, o->in1, o->in2);
2741 static ExitStatus op_srl(DisasContext *s, DisasOps *o)
2743 tcg_gen_shr_i64(o->out, o->in1, o->in2);
2747 static ExitStatus op_sfpc(DisasContext *s, DisasOps *o)
2749 gen_helper_sfpc(cpu_env, o->in2);
2753 #ifndef CONFIG_USER_ONLY
2754 static ExitStatus op_spka(DisasContext *s, DisasOps *o)
2756 check_privileged(s);
2757 tcg_gen_shri_i64(o->in2, o->in2, 4);
2758 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, PSW_SHIFT_KEY - 4, 4);
2762 static ExitStatus op_sske(DisasContext *s, DisasOps *o)
2764 check_privileged(s);
2765 gen_helper_sske(cpu_env, o->in1, o->in2);
2769 static ExitStatus op_ssm(DisasContext *s, DisasOps *o)
2771 check_privileged(s);
2772 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, 56, 8);
2776 static ExitStatus op_stap(DisasContext *s, DisasOps *o)
2778 check_privileged(s);
2779 /* ??? Surely cpu address != cpu number. In any case the previous
2780 version of this stored more than the required half-word, so it
2781 is unlikely this has ever been tested. */
2782 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
2786 static ExitStatus op_stck(DisasContext *s, DisasOps *o)
2788 gen_helper_stck(o->out, cpu_env);
2789 /* ??? We don't implement clock states. */
2790 gen_op_movi_cc(s, 0);
2794 static ExitStatus op_stcke(DisasContext *s, DisasOps *o)
2796 TCGv_i64 c1 = tcg_temp_new_i64();
2797 TCGv_i64 c2 = tcg_temp_new_i64();
2798 gen_helper_stck(c1, cpu_env);
2799 /* Shift the 64-bit value into its place as a zero-extended
2800 104-bit value. Note that "bit positions 64-103 are always
2801 non-zero so that they compare differently to STCK"; we set
2802 the least significant bit to 1. */
2803 tcg_gen_shli_i64(c2, c1, 56);
2804 tcg_gen_shri_i64(c1, c1, 8);
2805 tcg_gen_ori_i64(c2, c2, 0x10000);
2806 tcg_gen_qemu_st64(c1, o->in2, get_mem_index(s));
2807 tcg_gen_addi_i64(o->in2, o->in2, 8);
2808 tcg_gen_qemu_st64(c2, o->in2, get_mem_index(s));
2809 tcg_temp_free_i64(c1);
2810 tcg_temp_free_i64(c2);
2811 /* ??? We don't implement clock states. */
2812 gen_op_movi_cc(s, 0);
2816 static ExitStatus op_sckc(DisasContext *s, DisasOps *o)
2818 check_privileged(s);
2819 gen_helper_sckc(cpu_env, o->in2);
2823 static ExitStatus op_stckc(DisasContext *s, DisasOps *o)
2825 check_privileged(s);
2826 gen_helper_stckc(o->out, cpu_env);
2830 static ExitStatus op_stctg(DisasContext *s, DisasOps *o)
2832 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2833 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2834 check_privileged(s);
2835 potential_page_fault(s);
2836 gen_helper_stctg(cpu_env, r1, o->in2, r3);
2837 tcg_temp_free_i32(r1);
2838 tcg_temp_free_i32(r3);
2842 static ExitStatus op_stctl(DisasContext *s, DisasOps *o)
2844 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2845 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2846 check_privileged(s);
2847 potential_page_fault(s);
2848 gen_helper_stctl(cpu_env, r1, o->in2, r3);
2849 tcg_temp_free_i32(r1);
2850 tcg_temp_free_i32(r3);
2854 static ExitStatus op_stidp(DisasContext *s, DisasOps *o)
2856 check_privileged(s);
2857 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
2861 static ExitStatus op_spt(DisasContext *s, DisasOps *o)
2863 check_privileged(s);
2864 gen_helper_spt(cpu_env, o->in2);
2868 static ExitStatus op_stfl(DisasContext *s, DisasOps *o)
2871 /* We really ought to have more complete indication of facilities
2872 that we implement. Address this when STFLE is implemented. */
2873 check_privileged(s);
2874 f = tcg_const_i64(0xc0000000);
2875 a = tcg_const_i64(200);
2876 tcg_gen_qemu_st32(f, a, get_mem_index(s));
2877 tcg_temp_free_i64(f);
2878 tcg_temp_free_i64(a);
2882 static ExitStatus op_stpt(DisasContext *s, DisasOps *o)
2884 check_privileged(s);
2885 gen_helper_stpt(o->out, cpu_env);
2889 static ExitStatus op_stsi(DisasContext *s, DisasOps *o)
2891 check_privileged(s);
2892 potential_page_fault(s);
2893 gen_helper_stsi(cc_op, cpu_env, o->in2, regs[0], regs[1]);
2898 static ExitStatus op_spx(DisasContext *s, DisasOps *o)
2900 check_privileged(s);
2901 gen_helper_spx(cpu_env, o->in2);
2905 static ExitStatus op_subchannel(DisasContext *s, DisasOps *o)
2907 check_privileged(s);
2908 /* Not operational. */
2909 gen_op_movi_cc(s, 3);
2913 static ExitStatus op_stpx(DisasContext *s, DisasOps *o)
2915 check_privileged(s);
2916 tcg_gen_ld_i64(o->out, cpu_env, offsetof(CPUS390XState, psa));
2917 tcg_gen_andi_i64(o->out, o->out, 0x7fffe000);
2921 static ExitStatus op_stnosm(DisasContext *s, DisasOps *o)
2923 uint64_t i2 = get_field(s->fields, i2);
2926 check_privileged(s);
2928 /* It is important to do what the instruction name says: STORE THEN.
2929 If we let the output hook perform the store then if we fault and
2930 restart, we'll have the wrong SYSTEM MASK in place. */
2931 t = tcg_temp_new_i64();
2932 tcg_gen_shri_i64(t, psw_mask, 56);
2933 tcg_gen_qemu_st8(t, o->addr1, get_mem_index(s));
2934 tcg_temp_free_i64(t);
2936 if (s->fields->op == 0xac) {
2937 tcg_gen_andi_i64(psw_mask, psw_mask,
2938 (i2 << 56) | 0x00ffffffffffffffull);
2940 tcg_gen_ori_i64(psw_mask, psw_mask, i2 << 56);
2945 static ExitStatus op_stura(DisasContext *s, DisasOps *o)
2947 check_privileged(s);
2948 potential_page_fault(s);
2949 gen_helper_stura(cpu_env, o->in2, o->in1);
2954 static ExitStatus op_st8(DisasContext *s, DisasOps *o)
2956 tcg_gen_qemu_st8(o->in1, o->in2, get_mem_index(s));
2960 static ExitStatus op_st16(DisasContext *s, DisasOps *o)
2962 tcg_gen_qemu_st16(o->in1, o->in2, get_mem_index(s));
2966 static ExitStatus op_st32(DisasContext *s, DisasOps *o)
2968 tcg_gen_qemu_st32(o->in1, o->in2, get_mem_index(s));
2972 static ExitStatus op_st64(DisasContext *s, DisasOps *o)
2974 tcg_gen_qemu_st64(o->in1, o->in2, get_mem_index(s));
2978 static ExitStatus op_stam(DisasContext *s, DisasOps *o)
2980 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2981 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2982 potential_page_fault(s);
2983 gen_helper_stam(cpu_env, r1, o->in2, r3);
2984 tcg_temp_free_i32(r1);
2985 tcg_temp_free_i32(r3);
2989 static ExitStatus op_stcm(DisasContext *s, DisasOps *o)
2991 int m3 = get_field(s->fields, m3);
2992 int pos, base = s->insn->data;
2993 TCGv_i64 tmp = tcg_temp_new_i64();
2995 pos = base + ctz32(m3) * 8;
2998 /* Effectively a 32-bit store. */
2999 tcg_gen_shri_i64(tmp, o->in1, pos);
3000 tcg_gen_qemu_st32(tmp, o->in2, get_mem_index(s));
3006 /* Effectively a 16-bit store. */
3007 tcg_gen_shri_i64(tmp, o->in1, pos);
3008 tcg_gen_qemu_st16(tmp, o->in2, get_mem_index(s));
3015 /* Effectively an 8-bit store. */
3016 tcg_gen_shri_i64(tmp, o->in1, pos);
3017 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3021 /* This is going to be a sequence of shifts and stores. */
3022 pos = base + 32 - 8;
3025 tcg_gen_shri_i64(tmp, o->in1, pos);
3026 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3027 tcg_gen_addi_i64(o->in2, o->in2, 1);
3029 m3 = (m3 << 1) & 0xf;
3034 tcg_temp_free_i64(tmp);
3038 static ExitStatus op_stm(DisasContext *s, DisasOps *o)
3040 int r1 = get_field(s->fields, r1);
3041 int r3 = get_field(s->fields, r3);
3042 int size = s->insn->data;
3043 TCGv_i64 tsize = tcg_const_i64(size);
3047 tcg_gen_qemu_st64(regs[r1], o->in2, get_mem_index(s));
3049 tcg_gen_qemu_st32(regs[r1], o->in2, get_mem_index(s));
3054 tcg_gen_add_i64(o->in2, o->in2, tsize);
3058 tcg_temp_free_i64(tsize);
3062 static ExitStatus op_stmh(DisasContext *s, DisasOps *o)
3064 int r1 = get_field(s->fields, r1);
3065 int r3 = get_field(s->fields, r3);
3066 TCGv_i64 t = tcg_temp_new_i64();
3067 TCGv_i64 t4 = tcg_const_i64(4);
3068 TCGv_i64 t32 = tcg_const_i64(32);
3071 tcg_gen_shl_i64(t, regs[r1], t32);
3072 tcg_gen_qemu_st32(t, o->in2, get_mem_index(s));
3076 tcg_gen_add_i64(o->in2, o->in2, t4);
3080 tcg_temp_free_i64(t);
3081 tcg_temp_free_i64(t4);
3082 tcg_temp_free_i64(t32);
3086 static ExitStatus op_srst(DisasContext *s, DisasOps *o)
3088 potential_page_fault(s);
3089 gen_helper_srst(o->in1, cpu_env, regs[0], o->in1, o->in2);
3091 return_low128(o->in2);
3095 static ExitStatus op_sub(DisasContext *s, DisasOps *o)
3097 tcg_gen_sub_i64(o->out, o->in1, o->in2);
3101 static ExitStatus op_subb(DisasContext *s, DisasOps *o)
3106 tcg_gen_not_i64(o->in2, o->in2);
3107 tcg_gen_add_i64(o->out, o->in1, o->in2);
3109 /* XXX possible optimization point */
3111 cc = tcg_temp_new_i64();
3112 tcg_gen_extu_i32_i64(cc, cc_op);
3113 tcg_gen_shri_i64(cc, cc, 1);
3114 tcg_gen_add_i64(o->out, o->out, cc);
3115 tcg_temp_free_i64(cc);
3119 static ExitStatus op_svc(DisasContext *s, DisasOps *o)
3126 t = tcg_const_i32(get_field(s->fields, i1) & 0xff);
3127 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_code));
3128 tcg_temp_free_i32(t);
3130 t = tcg_const_i32(s->next_pc - s->pc);
3131 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_ilen));
3132 tcg_temp_free_i32(t);
3134 gen_exception(EXCP_SVC);
3135 return EXIT_NORETURN;
3138 static ExitStatus op_tceb(DisasContext *s, DisasOps *o)
3140 gen_helper_tceb(cc_op, o->in1, o->in2);
3145 static ExitStatus op_tcdb(DisasContext *s, DisasOps *o)
3147 gen_helper_tcdb(cc_op, o->in1, o->in2);
3152 static ExitStatus op_tcxb(DisasContext *s, DisasOps *o)
3154 gen_helper_tcxb(cc_op, o->out, o->out2, o->in2);
3159 #ifndef CONFIG_USER_ONLY
3160 static ExitStatus op_tprot(DisasContext *s, DisasOps *o)
3162 potential_page_fault(s);
3163 gen_helper_tprot(cc_op, o->addr1, o->in2);
3169 static ExitStatus op_tr(DisasContext *s, DisasOps *o)
3171 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3172 potential_page_fault(s);
3173 gen_helper_tr(cpu_env, l, o->addr1, o->in2);
3174 tcg_temp_free_i32(l);
3179 static ExitStatus op_unpk(DisasContext *s, DisasOps *o)
3181 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3182 potential_page_fault(s);
3183 gen_helper_unpk(cpu_env, l, o->addr1, o->in2);
3184 tcg_temp_free_i32(l);
3188 static ExitStatus op_xc(DisasContext *s, DisasOps *o)
3190 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3191 potential_page_fault(s);
3192 gen_helper_xc(cc_op, cpu_env, l, o->addr1, o->in2);
3193 tcg_temp_free_i32(l);
3198 static ExitStatus op_xor(DisasContext *s, DisasOps *o)
3200 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3204 static ExitStatus op_xori(DisasContext *s, DisasOps *o)
3206 int shift = s->insn->data & 0xff;
3207 int size = s->insn->data >> 8;
3208 uint64_t mask = ((1ull << size) - 1) << shift;
3211 tcg_gen_shli_i64(o->in2, o->in2, shift);
3212 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3214 /* Produce the CC from only the bits manipulated. */
3215 tcg_gen_andi_i64(cc_dst, o->out, mask);
3216 set_cc_nz_u64(s, cc_dst);
3220 static ExitStatus op_zero(DisasContext *s, DisasOps *o)
3222 o->out = tcg_const_i64(0);
3226 static ExitStatus op_zero2(DisasContext *s, DisasOps *o)
3228 o->out = tcg_const_i64(0);
3234 /* ====================================================================== */
3235 /* The "Cc OUTput" generators. Given the generated output (and in some cases
3236 the original inputs), update the various cc data structures in order to
3237 be able to compute the new condition code. */
3239 static void cout_abs32(DisasContext *s, DisasOps *o)
3241 gen_op_update1_cc_i64(s, CC_OP_ABS_32, o->out);
3244 static void cout_abs64(DisasContext *s, DisasOps *o)
3246 gen_op_update1_cc_i64(s, CC_OP_ABS_64, o->out);
3249 static void cout_adds32(DisasContext *s, DisasOps *o)
3251 gen_op_update3_cc_i64(s, CC_OP_ADD_32, o->in1, o->in2, o->out);
3254 static void cout_adds64(DisasContext *s, DisasOps *o)
3256 gen_op_update3_cc_i64(s, CC_OP_ADD_64, o->in1, o->in2, o->out);
3259 static void cout_addu32(DisasContext *s, DisasOps *o)
3261 gen_op_update3_cc_i64(s, CC_OP_ADDU_32, o->in1, o->in2, o->out);
3264 static void cout_addu64(DisasContext *s, DisasOps *o)
3266 gen_op_update3_cc_i64(s, CC_OP_ADDU_64, o->in1, o->in2, o->out);
3269 static void cout_addc32(DisasContext *s, DisasOps *o)
3271 gen_op_update3_cc_i64(s, CC_OP_ADDC_32, o->in1, o->in2, o->out);
3274 static void cout_addc64(DisasContext *s, DisasOps *o)
3276 gen_op_update3_cc_i64(s, CC_OP_ADDC_64, o->in1, o->in2, o->out);
3279 static void cout_cmps32(DisasContext *s, DisasOps *o)
3281 gen_op_update2_cc_i64(s, CC_OP_LTGT_32, o->in1, o->in2);
3284 static void cout_cmps64(DisasContext *s, DisasOps *o)
3286 gen_op_update2_cc_i64(s, CC_OP_LTGT_64, o->in1, o->in2);
3289 static void cout_cmpu32(DisasContext *s, DisasOps *o)
3291 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_32, o->in1, o->in2);
3294 static void cout_cmpu64(DisasContext *s, DisasOps *o)
3296 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, o->in1, o->in2);
3299 static void cout_f32(DisasContext *s, DisasOps *o)
3301 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, o->out);
3304 static void cout_f64(DisasContext *s, DisasOps *o)
3306 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, o->out);
3309 static void cout_f128(DisasContext *s, DisasOps *o)
3311 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, o->out, o->out2);
3314 static void cout_nabs32(DisasContext *s, DisasOps *o)
3316 gen_op_update1_cc_i64(s, CC_OP_NABS_32, o->out);
3319 static void cout_nabs64(DisasContext *s, DisasOps *o)
3321 gen_op_update1_cc_i64(s, CC_OP_NABS_64, o->out);
3324 static void cout_neg32(DisasContext *s, DisasOps *o)
3326 gen_op_update1_cc_i64(s, CC_OP_COMP_32, o->out);
3329 static void cout_neg64(DisasContext *s, DisasOps *o)
3331 gen_op_update1_cc_i64(s, CC_OP_COMP_64, o->out);
3334 static void cout_nz32(DisasContext *s, DisasOps *o)
3336 tcg_gen_ext32u_i64(cc_dst, o->out);
3337 gen_op_update1_cc_i64(s, CC_OP_NZ, cc_dst);
3340 static void cout_nz64(DisasContext *s, DisasOps *o)
3342 gen_op_update1_cc_i64(s, CC_OP_NZ, o->out);
3345 static void cout_s32(DisasContext *s, DisasOps *o)
3347 gen_op_update1_cc_i64(s, CC_OP_LTGT0_32, o->out);
3350 static void cout_s64(DisasContext *s, DisasOps *o)
3352 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, o->out);
3355 static void cout_subs32(DisasContext *s, DisasOps *o)
3357 gen_op_update3_cc_i64(s, CC_OP_SUB_32, o->in1, o->in2, o->out);
3360 static void cout_subs64(DisasContext *s, DisasOps *o)
3362 gen_op_update3_cc_i64(s, CC_OP_SUB_64, o->in1, o->in2, o->out);
3365 static void cout_subu32(DisasContext *s, DisasOps *o)
3367 gen_op_update3_cc_i64(s, CC_OP_SUBU_32, o->in1, o->in2, o->out);
3370 static void cout_subu64(DisasContext *s, DisasOps *o)
3372 gen_op_update3_cc_i64(s, CC_OP_SUBU_64, o->in1, o->in2, o->out);
3375 static void cout_subb32(DisasContext *s, DisasOps *o)
3377 gen_op_update3_cc_i64(s, CC_OP_SUBB_32, o->in1, o->in2, o->out);
3380 static void cout_subb64(DisasContext *s, DisasOps *o)
3382 gen_op_update3_cc_i64(s, CC_OP_SUBB_64, o->in1, o->in2, o->out);
3385 static void cout_tm32(DisasContext *s, DisasOps *o)
3387 gen_op_update2_cc_i64(s, CC_OP_TM_32, o->in1, o->in2);
3390 static void cout_tm64(DisasContext *s, DisasOps *o)
3392 gen_op_update2_cc_i64(s, CC_OP_TM_64, o->in1, o->in2);
3395 /* ====================================================================== */
3396 /* The "PREPeration" generators. These initialize the DisasOps.OUT fields
3397 with the TCG register to which we will write. Used in combination with
3398 the "wout" generators, in some cases we need a new temporary, and in
3399 some cases we can write to a TCG global. */
3401 static void prep_new(DisasContext *s, DisasFields *f, DisasOps *o)
3403 o->out = tcg_temp_new_i64();
3406 static void prep_new_P(DisasContext *s, DisasFields *f, DisasOps *o)
3408 o->out = tcg_temp_new_i64();
3409 o->out2 = tcg_temp_new_i64();
3412 static void prep_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3414 o->out = regs[get_field(f, r1)];
3418 static void prep_r1_P(DisasContext *s, DisasFields *f, DisasOps *o)
3420 /* ??? Specification exception: r1 must be even. */
3421 int r1 = get_field(f, r1);
3423 o->out2 = regs[(r1 + 1) & 15];
3424 o->g_out = o->g_out2 = true;
3427 static void prep_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3429 o->out = fregs[get_field(f, r1)];
3433 static void prep_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3435 /* ??? Specification exception: r1 must be < 14. */
3436 int r1 = get_field(f, r1);
3438 o->out2 = fregs[(r1 + 2) & 15];
3439 o->g_out = o->g_out2 = true;
3442 /* ====================================================================== */
3443 /* The "Write OUTput" generators. These generally perform some non-trivial
3444 copy of data to TCG globals, or to main memory. The trivial cases are
3445 generally handled by having a "prep" generator install the TCG global
3446 as the destination of the operation. */
3448 static void wout_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3450 store_reg(get_field(f, r1), o->out);
3453 static void wout_r1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3455 int r1 = get_field(f, r1);
3456 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 8);
3459 static void wout_r1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3461 int r1 = get_field(f, r1);
3462 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 16);
3465 static void wout_r1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3467 store_reg32_i64(get_field(f, r1), o->out);
3470 static void wout_r1_P32(DisasContext *s, DisasFields *f, DisasOps *o)
3472 /* ??? Specification exception: r1 must be even. */
3473 int r1 = get_field(f, r1);
3474 store_reg32_i64(r1, o->out);
3475 store_reg32_i64((r1 + 1) & 15, o->out2);
3478 static void wout_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3480 /* ??? Specification exception: r1 must be even. */
3481 int r1 = get_field(f, r1);
3482 store_reg32_i64((r1 + 1) & 15, o->out);
3483 tcg_gen_shri_i64(o->out, o->out, 32);
3484 store_reg32_i64(r1, o->out);
3487 static void wout_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3489 store_freg32_i64(get_field(f, r1), o->out);
3492 static void wout_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3494 store_freg(get_field(f, r1), o->out);
3497 static void wout_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3499 /* ??? Specification exception: r1 must be < 14. */
3500 int f1 = get_field(s->fields, r1);
3501 store_freg(f1, o->out);
3502 store_freg((f1 + 2) & 15, o->out2);
3505 static void wout_cond_r1r2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3507 if (get_field(f, r1) != get_field(f, r2)) {
3508 store_reg32_i64(get_field(f, r1), o->out);
3512 static void wout_cond_e1e2(DisasContext *s, DisasFields *f, DisasOps *o)
3514 if (get_field(f, r1) != get_field(f, r2)) {
3515 store_freg32_i64(get_field(f, r1), o->out);
3519 static void wout_m1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3521 tcg_gen_qemu_st8(o->out, o->addr1, get_mem_index(s));
3524 static void wout_m1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3526 tcg_gen_qemu_st16(o->out, o->addr1, get_mem_index(s));
3529 static void wout_m1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3531 tcg_gen_qemu_st32(o->out, o->addr1, get_mem_index(s));
3534 static void wout_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3536 tcg_gen_qemu_st64(o->out, o->addr1, get_mem_index(s));
3539 static void wout_m2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3541 tcg_gen_qemu_st32(o->out, o->in2, get_mem_index(s));
3544 /* ====================================================================== */
3545 /* The "INput 1" generators. These load the first operand to an insn. */
3547 static void in1_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3549 o->in1 = load_reg(get_field(f, r1));
3552 static void in1_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3554 o->in1 = regs[get_field(f, r1)];
3558 static void in1_r1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3560 o->in1 = tcg_temp_new_i64();
3561 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r1)]);
3564 static void in1_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3566 o->in1 = tcg_temp_new_i64();
3567 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r1)]);
3570 static void in1_r1_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
3572 o->in1 = tcg_temp_new_i64();
3573 tcg_gen_shri_i64(o->in1, regs[get_field(f, r1)], 32);
3576 static void in1_r1p1(DisasContext *s, DisasFields *f, DisasOps *o)
3578 /* ??? Specification exception: r1 must be even. */
3579 int r1 = get_field(f, r1);
3580 o->in1 = load_reg((r1 + 1) & 15);
3583 static void in1_r1p1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3585 /* ??? Specification exception: r1 must be even. */
3586 int r1 = get_field(f, r1);
3587 o->in1 = tcg_temp_new_i64();
3588 tcg_gen_ext32s_i64(o->in1, regs[(r1 + 1) & 15]);
3591 static void in1_r1p1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3593 /* ??? Specification exception: r1 must be even. */
3594 int r1 = get_field(f, r1);
3595 o->in1 = tcg_temp_new_i64();
3596 tcg_gen_ext32u_i64(o->in1, regs[(r1 + 1) & 15]);
3599 static void in1_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3601 /* ??? Specification exception: r1 must be even. */
3602 int r1 = get_field(f, r1);
3603 o->in1 = tcg_temp_new_i64();
3604 tcg_gen_concat32_i64(o->in1, regs[r1 + 1], regs[r1]);
3607 static void in1_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3609 o->in1 = load_reg(get_field(f, r2));
3612 static void in1_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3614 o->in1 = load_reg(get_field(f, r3));
3617 static void in1_r3_o(DisasContext *s, DisasFields *f, DisasOps *o)
3619 o->in1 = regs[get_field(f, r3)];
3623 static void in1_r3_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3625 o->in1 = tcg_temp_new_i64();
3626 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r3)]);
3629 static void in1_r3_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3631 o->in1 = tcg_temp_new_i64();
3632 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r3)]);
3635 static void in1_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3637 o->in1 = load_freg32_i64(get_field(f, r1));
3640 static void in1_f1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3642 o->in1 = fregs[get_field(f, r1)];
3646 static void in1_x1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3648 /* ??? Specification exception: r1 must be < 14. */
3649 int r1 = get_field(f, r1);
3651 o->out2 = fregs[(r1 + 2) & 15];
3652 o->g_out = o->g_out2 = true;
3655 static void in1_la1(DisasContext *s, DisasFields *f, DisasOps *o)
3657 o->addr1 = get_address(s, 0, get_field(f, b1), get_field(f, d1));
3660 static void in1_la2(DisasContext *s, DisasFields *f, DisasOps *o)
3662 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3663 o->addr1 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3666 static void in1_m1_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3669 o->in1 = tcg_temp_new_i64();
3670 tcg_gen_qemu_ld8u(o->in1, o->addr1, get_mem_index(s));
3673 static void in1_m1_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3676 o->in1 = tcg_temp_new_i64();
3677 tcg_gen_qemu_ld16s(o->in1, o->addr1, get_mem_index(s));
3680 static void in1_m1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3683 o->in1 = tcg_temp_new_i64();
3684 tcg_gen_qemu_ld16u(o->in1, o->addr1, get_mem_index(s));
3687 static void in1_m1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3690 o->in1 = tcg_temp_new_i64();
3691 tcg_gen_qemu_ld32s(o->in1, o->addr1, get_mem_index(s));
3694 static void in1_m1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3697 o->in1 = tcg_temp_new_i64();
3698 tcg_gen_qemu_ld32u(o->in1, o->addr1, get_mem_index(s));
3701 static void in1_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3704 o->in1 = tcg_temp_new_i64();
3705 tcg_gen_qemu_ld64(o->in1, o->addr1, get_mem_index(s));
3708 /* ====================================================================== */
3709 /* The "INput 2" generators. These load the second operand to an insn. */
3711 static void in2_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3713 o->in2 = regs[get_field(f, r1)];
3717 static void in2_r1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3719 o->in2 = tcg_temp_new_i64();
3720 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r1)]);
3723 static void in2_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3725 o->in2 = tcg_temp_new_i64();
3726 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r1)]);
3729 static void in2_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3731 o->in2 = load_reg(get_field(f, r2));
3734 static void in2_r2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3736 o->in2 = regs[get_field(f, r2)];
3740 static void in2_r2_nz(DisasContext *s, DisasFields *f, DisasOps *o)
3742 int r2 = get_field(f, r2);
3744 o->in2 = load_reg(r2);
3748 static void in2_r2_8s(DisasContext *s, DisasFields *f, DisasOps *o)
3750 o->in2 = tcg_temp_new_i64();
3751 tcg_gen_ext8s_i64(o->in2, regs[get_field(f, r2)]);
3754 static void in2_r2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3756 o->in2 = tcg_temp_new_i64();
3757 tcg_gen_ext8u_i64(o->in2, regs[get_field(f, r2)]);
3760 static void in2_r2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3762 o->in2 = tcg_temp_new_i64();
3763 tcg_gen_ext16s_i64(o->in2, regs[get_field(f, r2)]);
3766 static void in2_r2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3768 o->in2 = tcg_temp_new_i64();
3769 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r2)]);
3772 static void in2_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3774 o->in2 = load_reg(get_field(f, r3));
3777 static void in2_r2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3779 o->in2 = tcg_temp_new_i64();
3780 tcg_gen_ext32s_i64(o->in2, regs[get_field(f, r2)]);
3783 static void in2_r2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3785 o->in2 = tcg_temp_new_i64();
3786 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r2)]);
3789 static void in2_e2(DisasContext *s, DisasFields *f, DisasOps *o)
3791 o->in2 = load_freg32_i64(get_field(f, r2));
3794 static void in2_f2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3796 o->in2 = fregs[get_field(f, r2)];
3800 static void in2_x2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3802 /* ??? Specification exception: r1 must be < 14. */
3803 int r2 = get_field(f, r2);
3805 o->in2 = fregs[(r2 + 2) & 15];
3806 o->g_in1 = o->g_in2 = true;
3809 static void in2_ra2(DisasContext *s, DisasFields *f, DisasOps *o)
3811 o->in2 = get_address(s, 0, get_field(f, r2), 0);
3814 static void in2_a2(DisasContext *s, DisasFields *f, DisasOps *o)
3816 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3817 o->in2 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3820 static void in2_ri2(DisasContext *s, DisasFields *f, DisasOps *o)
3822 o->in2 = tcg_const_i64(s->pc + (int64_t)get_field(f, i2) * 2);
3825 static void in2_sh32(DisasContext *s, DisasFields *f, DisasOps *o)
3827 help_l2_shift(s, f, o, 31);
3830 static void in2_sh64(DisasContext *s, DisasFields *f, DisasOps *o)
3832 help_l2_shift(s, f, o, 63);
3835 static void in2_m2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3838 tcg_gen_qemu_ld8u(o->in2, o->in2, get_mem_index(s));
3841 static void in2_m2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3844 tcg_gen_qemu_ld16s(o->in2, o->in2, get_mem_index(s));
3847 static void in2_m2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3850 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
3853 static void in2_m2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3856 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
3859 static void in2_m2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3862 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
3865 static void in2_m2_64(DisasContext *s, DisasFields *f, DisasOps *o)
3868 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
3871 static void in2_mri2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3874 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
3877 static void in2_mri2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3880 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
3883 static void in2_mri2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3886 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
3889 static void in2_mri2_64(DisasContext *s, DisasFields *f, DisasOps *o)
3892 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
3895 static void in2_i2(DisasContext *s, DisasFields *f, DisasOps *o)
3897 o->in2 = tcg_const_i64(get_field(f, i2));
3900 static void in2_i2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3902 o->in2 = tcg_const_i64((uint8_t)get_field(f, i2));
3905 static void in2_i2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3907 o->in2 = tcg_const_i64((uint16_t)get_field(f, i2));
3910 static void in2_i2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3912 o->in2 = tcg_const_i64((uint32_t)get_field(f, i2));
3915 static void in2_i2_16u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
3917 uint64_t i2 = (uint16_t)get_field(f, i2);
3918 o->in2 = tcg_const_i64(i2 << s->insn->data);
3921 static void in2_i2_32u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
3923 uint64_t i2 = (uint32_t)get_field(f, i2);
3924 o->in2 = tcg_const_i64(i2 << s->insn->data);
3927 /* ====================================================================== */
3929 /* Find opc within the table of insns. This is formulated as a switch
3930 statement so that (1) we get compile-time notice of cut-paste errors
3931 for duplicated opcodes, and (2) the compiler generates the binary
3932 search tree, rather than us having to post-process the table. */
3934 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
3935 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
3937 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
3939 enum DisasInsnEnum {
3940 #include "insn-data.def"
3944 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
3949 .help_in1 = in1_##I1, \
3950 .help_in2 = in2_##I2, \
3951 .help_prep = prep_##P, \
3952 .help_wout = wout_##W, \
3953 .help_cout = cout_##CC, \
3954 .help_op = op_##OP, \
3958 /* Allow 0 to be used for NULL in the table below. */
3966 static const DisasInsn insn_info[] = {
3967 #include "insn-data.def"
3971 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
3972 case OPC: return &insn_info[insn_ ## NM];
3974 static const DisasInsn *lookup_opc(uint16_t opc)
3977 #include "insn-data.def"
3986 /* Extract a field from the insn. The INSN should be left-aligned in
3987 the uint64_t so that we can more easily utilize the big-bit-endian
3988 definitions we extract from the Principals of Operation. */
3990 static void extract_field(DisasFields *o, const DisasField *f, uint64_t insn)
3998 /* Zero extract the field from the insn. */
3999 r = (insn << f->beg) >> (64 - f->size);
4001 /* Sign-extend, or un-swap the field as necessary. */
4003 case 0: /* unsigned */
4005 case 1: /* signed */
4006 assert(f->size <= 32);
4007 m = 1u << (f->size - 1);
4010 case 2: /* dl+dh split, signed 20 bit. */
4011 r = ((int8_t)r << 12) | (r >> 8);
4017 /* Validate that the "compressed" encoding we selected above is valid.
4018 I.e. we havn't make two different original fields overlap. */
4019 assert(((o->presentC >> f->indexC) & 1) == 0);
4020 o->presentC |= 1 << f->indexC;
4021 o->presentO |= 1 << f->indexO;
4023 o->c[f->indexC] = r;
4026 /* Lookup the insn at the current PC, extracting the operands into O and
4027 returning the info struct for the insn. Returns NULL for invalid insn. */
4029 static const DisasInsn *extract_insn(CPUS390XState *env, DisasContext *s,
4032 uint64_t insn, pc = s->pc;
4034 const DisasInsn *info;
4036 insn = ld_code2(env, pc);
4037 op = (insn >> 8) & 0xff;
4038 ilen = get_ilen(op);
4039 s->next_pc = s->pc + ilen;
4046 insn = ld_code4(env, pc) << 32;
4049 insn = (insn << 48) | (ld_code4(env, pc + 2) << 16);
4055 /* We can't actually determine the insn format until we've looked up
4056 the full insn opcode. Which we can't do without locating the
4057 secondary opcode. Assume by default that OP2 is at bit 40; for
4058 those smaller insns that don't actually have a secondary opcode
4059 this will correctly result in OP2 = 0. */
4065 case 0xb2: /* S, RRF, RRE */
4066 case 0xb3: /* RRE, RRD, RRF */
4067 case 0xb9: /* RRE, RRF */
4068 case 0xe5: /* SSE, SIL */
4069 op2 = (insn << 8) >> 56;
4073 case 0xc0: /* RIL */
4074 case 0xc2: /* RIL */
4075 case 0xc4: /* RIL */
4076 case 0xc6: /* RIL */
4077 case 0xc8: /* SSF */
4078 case 0xcc: /* RIL */
4079 op2 = (insn << 12) >> 60;
4081 case 0xd0 ... 0xdf: /* SS */
4087 case 0xee ... 0xf3: /* SS */
4088 case 0xf8 ... 0xfd: /* SS */
4092 op2 = (insn << 40) >> 56;
4096 memset(f, 0, sizeof(*f));
4100 /* Lookup the instruction. */
4101 info = lookup_opc(op << 8 | op2);
4103 /* If we found it, extract the operands. */
4105 DisasFormat fmt = info->fmt;
4108 for (i = 0; i < NUM_C_FIELD; ++i) {
4109 extract_field(f, &format_info[fmt].op[i], insn);
4115 static ExitStatus translate_one(CPUS390XState *env, DisasContext *s)
4117 const DisasInsn *insn;
4118 ExitStatus ret = NO_EXIT;
4122 insn = extract_insn(env, s, &f);
4124 /* If not found, try the old interpreter. This includes ILLOPC. */
4126 disas_s390_insn(env, s);
4127 switch (s->is_jmp) {
4135 ret = EXIT_PC_UPDATED;
4138 ret = EXIT_NORETURN;
4148 /* Set up the strutures we use to communicate with the helpers. */
4151 o.g_out = o.g_out2 = o.g_in1 = o.g_in2 = false;
4152 TCGV_UNUSED_I64(o.out);
4153 TCGV_UNUSED_I64(o.out2);
4154 TCGV_UNUSED_I64(o.in1);
4155 TCGV_UNUSED_I64(o.in2);
4156 TCGV_UNUSED_I64(o.addr1);
4158 /* Implement the instruction. */
4159 if (insn->help_in1) {
4160 insn->help_in1(s, &f, &o);
4162 if (insn->help_in2) {
4163 insn->help_in2(s, &f, &o);
4165 if (insn->help_prep) {
4166 insn->help_prep(s, &f, &o);
4168 if (insn->help_op) {
4169 ret = insn->help_op(s, &o);
4171 if (insn->help_wout) {
4172 insn->help_wout(s, &f, &o);
4174 if (insn->help_cout) {
4175 insn->help_cout(s, &o);
4178 /* Free any temporaries created by the helpers. */
4179 if (!TCGV_IS_UNUSED_I64(o.out) && !o.g_out) {
4180 tcg_temp_free_i64(o.out);
4182 if (!TCGV_IS_UNUSED_I64(o.out2) && !o.g_out2) {
4183 tcg_temp_free_i64(o.out2);
4185 if (!TCGV_IS_UNUSED_I64(o.in1) && !o.g_in1) {
4186 tcg_temp_free_i64(o.in1);
4188 if (!TCGV_IS_UNUSED_I64(o.in2) && !o.g_in2) {
4189 tcg_temp_free_i64(o.in2);
4191 if (!TCGV_IS_UNUSED_I64(o.addr1)) {
4192 tcg_temp_free_i64(o.addr1);
4195 /* Advance to the next instruction. */
4200 static inline void gen_intermediate_code_internal(CPUS390XState *env,
4201 TranslationBlock *tb,
4205 target_ulong pc_start;
4206 uint64_t next_page_start;
4207 uint16_t *gen_opc_end;
4209 int num_insns, max_insns;
4217 if (!(tb->flags & FLAG_MASK_64)) {
4218 pc_start &= 0x7fffffff;
4223 dc.cc_op = CC_OP_DYNAMIC;
4224 do_debug = dc.singlestep_enabled = env->singlestep_enabled;
4225 dc.is_jmp = DISAS_NEXT;
4227 gen_opc_end = tcg_ctx.gen_opc_buf + OPC_MAX_SIZE;
4229 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
4232 max_insns = tb->cflags & CF_COUNT_MASK;
4233 if (max_insns == 0) {
4234 max_insns = CF_COUNT_MASK;
4241 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4245 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4248 tcg_ctx.gen_opc_pc[lj] = dc.pc;
4249 gen_opc_cc_op[lj] = dc.cc_op;
4250 tcg_ctx.gen_opc_instr_start[lj] = 1;
4251 tcg_ctx.gen_opc_icount[lj] = num_insns;
4253 if (++num_insns == max_insns && (tb->cflags & CF_LAST_IO)) {
4257 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) {
4258 tcg_gen_debug_insn_start(dc.pc);
4262 if (unlikely(!QTAILQ_EMPTY(&env->breakpoints))) {
4263 QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
4264 if (bp->pc == dc.pc) {
4265 status = EXIT_PC_STALE;
4271 if (status == NO_EXIT) {
4272 status = translate_one(env, &dc);
4275 /* If we reach a page boundary, are single stepping,
4276 or exhaust instruction count, stop generation. */
4277 if (status == NO_EXIT
4278 && (dc.pc >= next_page_start
4279 || tcg_ctx.gen_opc_ptr >= gen_opc_end
4280 || num_insns >= max_insns
4282 || env->singlestep_enabled)) {
4283 status = EXIT_PC_STALE;
4285 } while (status == NO_EXIT);
4287 if (tb->cflags & CF_LAST_IO) {
4296 update_psw_addr(&dc);
4298 case EXIT_PC_UPDATED:
4299 if (singlestep && dc.cc_op != CC_OP_DYNAMIC) {
4300 gen_op_calc_cc(&dc);
4302 /* Next TB starts off with CC_OP_DYNAMIC,
4303 so make sure the cc op type is in env */
4304 gen_op_set_cc_op(&dc);
4307 gen_exception(EXCP_DEBUG);
4309 /* Generate the return instruction */
4317 gen_icount_end(tb, num_insns);
4318 *tcg_ctx.gen_opc_ptr = INDEX_op_end;
4320 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4323 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4326 tb->size = dc.pc - pc_start;
4327 tb->icount = num_insns;
4330 #if defined(S390X_DEBUG_DISAS)
4331 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
4332 qemu_log("IN: %s\n", lookup_symbol(pc_start));
4333 log_target_disas(env, pc_start, dc.pc - pc_start, 1);
4339 void gen_intermediate_code (CPUS390XState *env, struct TranslationBlock *tb)
4341 gen_intermediate_code_internal(env, tb, 0);
4344 void gen_intermediate_code_pc (CPUS390XState *env, struct TranslationBlock *tb)
4346 gen_intermediate_code_internal(env, tb, 1);
4349 void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb, int pc_pos)
4352 env->psw.addr = tcg_ctx.gen_opc_pc[pc_pos];
4353 cc_op = gen_opc_cc_op[pc_pos];
4354 if ((cc_op != CC_OP_DYNAMIC) && (cc_op != CC_OP_STATIC)) {