4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
28 # define LOG_DISAS(...) do { } while (0)
32 #include "disas/disas.h"
35 #include "qemu/host-utils.h"
37 /* global register indexes */
38 static TCGv_ptr cpu_env;
40 #include "exec/gen-icount.h"
46 /* Information that (most) every instruction needs to manipulate. */
47 typedef struct DisasContext DisasContext;
48 typedef struct DisasInsn DisasInsn;
49 typedef struct DisasFields DisasFields;
52 struct TranslationBlock *tb;
53 const DisasInsn *insn;
57 bool singlestep_enabled;
61 /* Information carried about a condition to be evaluated. */
68 struct { TCGv_i64 a, b; } s64;
69 struct { TCGv_i32 a, b; } s32;
75 static void gen_op_calc_cc(DisasContext *s);
77 #ifdef DEBUG_INLINE_BRANCHES
78 static uint64_t inline_branch_hit[CC_OP_MAX];
79 static uint64_t inline_branch_miss[CC_OP_MAX];
82 static inline void debug_insn(uint64_t insn)
84 LOG_DISAS("insn: 0x%" PRIx64 "\n", insn);
87 static inline uint64_t pc_to_link_info(DisasContext *s, uint64_t pc)
89 if (!(s->tb->flags & FLAG_MASK_64)) {
90 if (s->tb->flags & FLAG_MASK_32) {
91 return pc | 0x80000000;
97 void cpu_dump_state(CPUS390XState *env, FILE *f, fprintf_function cpu_fprintf,
102 if (env->cc_op > 3) {
103 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %15s\n",
104 env->psw.mask, env->psw.addr, cc_name(env->cc_op));
106 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %02x\n",
107 env->psw.mask, env->psw.addr, env->cc_op);
110 for (i = 0; i < 16; i++) {
111 cpu_fprintf(f, "R%02d=%016" PRIx64, i, env->regs[i]);
113 cpu_fprintf(f, "\n");
119 for (i = 0; i < 16; i++) {
120 cpu_fprintf(f, "F%02d=%016" PRIx64, i, env->fregs[i].ll);
122 cpu_fprintf(f, "\n");
128 #ifndef CONFIG_USER_ONLY
129 for (i = 0; i < 16; i++) {
130 cpu_fprintf(f, "C%02d=%016" PRIx64, i, env->cregs[i]);
132 cpu_fprintf(f, "\n");
139 #ifdef DEBUG_INLINE_BRANCHES
140 for (i = 0; i < CC_OP_MAX; i++) {
141 cpu_fprintf(f, " %15s = %10ld\t%10ld\n", cc_name(i),
142 inline_branch_miss[i], inline_branch_hit[i]);
146 cpu_fprintf(f, "\n");
149 static TCGv_i64 psw_addr;
150 static TCGv_i64 psw_mask;
152 static TCGv_i32 cc_op;
153 static TCGv_i64 cc_src;
154 static TCGv_i64 cc_dst;
155 static TCGv_i64 cc_vr;
157 static char cpu_reg_names[32][4];
158 static TCGv_i64 regs[16];
159 static TCGv_i64 fregs[16];
161 static uint8_t gen_opc_cc_op[OPC_BUF_SIZE];
163 void s390x_translate_init(void)
167 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
168 psw_addr = tcg_global_mem_new_i64(TCG_AREG0,
169 offsetof(CPUS390XState, psw.addr),
171 psw_mask = tcg_global_mem_new_i64(TCG_AREG0,
172 offsetof(CPUS390XState, psw.mask),
175 cc_op = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUS390XState, cc_op),
177 cc_src = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_src),
179 cc_dst = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_dst),
181 cc_vr = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_vr),
184 for (i = 0; i < 16; i++) {
185 snprintf(cpu_reg_names[i], sizeof(cpu_reg_names[0]), "r%d", i);
186 regs[i] = tcg_global_mem_new(TCG_AREG0,
187 offsetof(CPUS390XState, regs[i]),
191 for (i = 0; i < 16; i++) {
192 snprintf(cpu_reg_names[i + 16], sizeof(cpu_reg_names[0]), "f%d", i);
193 fregs[i] = tcg_global_mem_new(TCG_AREG0,
194 offsetof(CPUS390XState, fregs[i].d),
195 cpu_reg_names[i + 16]);
198 /* register helpers */
203 static inline TCGv_i64 load_reg(int reg)
205 TCGv_i64 r = tcg_temp_new_i64();
206 tcg_gen_mov_i64(r, regs[reg]);
210 static inline TCGv_i64 load_freg(int reg)
212 TCGv_i64 r = tcg_temp_new_i64();
213 tcg_gen_mov_i64(r, fregs[reg]);
217 static inline TCGv_i32 load_freg32(int reg)
219 TCGv_i32 r = tcg_temp_new_i32();
220 #if HOST_LONG_BITS == 32
221 tcg_gen_mov_i32(r, TCGV_HIGH(fregs[reg]));
223 tcg_gen_shri_i64(MAKE_TCGV_I64(GET_TCGV_I32(r)), fregs[reg], 32);
228 static inline TCGv_i64 load_freg32_i64(int reg)
230 TCGv_i64 r = tcg_temp_new_i64();
231 tcg_gen_shri_i64(r, fregs[reg], 32);
235 static inline TCGv_i32 load_reg32(int reg)
237 TCGv_i32 r = tcg_temp_new_i32();
238 tcg_gen_trunc_i64_i32(r, regs[reg]);
242 static inline TCGv_i64 load_reg32_i64(int reg)
244 TCGv_i64 r = tcg_temp_new_i64();
245 tcg_gen_ext32s_i64(r, regs[reg]);
249 static inline void store_reg(int reg, TCGv_i64 v)
251 tcg_gen_mov_i64(regs[reg], v);
254 static inline void store_freg(int reg, TCGv_i64 v)
256 tcg_gen_mov_i64(fregs[reg], v);
259 static inline void store_reg32(int reg, TCGv_i32 v)
261 /* 32 bit register writes keep the upper half */
262 #if HOST_LONG_BITS == 32
263 tcg_gen_mov_i32(TCGV_LOW(regs[reg]), v);
265 tcg_gen_deposit_i64(regs[reg], regs[reg],
266 MAKE_TCGV_I64(GET_TCGV_I32(v)), 0, 32);
270 static inline void store_reg32_i64(int reg, TCGv_i64 v)
272 /* 32 bit register writes keep the upper half */
273 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 0, 32);
276 static inline void store_reg32h_i64(int reg, TCGv_i64 v)
278 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 32, 32);
281 static inline void store_freg32(int reg, TCGv_i32 v)
283 /* 32 bit register writes keep the lower half */
284 #if HOST_LONG_BITS == 32
285 tcg_gen_mov_i32(TCGV_HIGH(fregs[reg]), v);
287 tcg_gen_deposit_i64(fregs[reg], fregs[reg],
288 MAKE_TCGV_I64(GET_TCGV_I32(v)), 32, 32);
292 static inline void store_freg32_i64(int reg, TCGv_i64 v)
294 tcg_gen_deposit_i64(fregs[reg], fregs[reg], v, 32, 32);
297 static inline void return_low128(TCGv_i64 dest)
299 tcg_gen_ld_i64(dest, cpu_env, offsetof(CPUS390XState, retxl));
302 static inline void update_psw_addr(DisasContext *s)
305 tcg_gen_movi_i64(psw_addr, s->pc);
308 static inline void potential_page_fault(DisasContext *s)
310 #ifndef CONFIG_USER_ONLY
316 static inline uint64_t ld_code2(CPUS390XState *env, uint64_t pc)
318 return (uint64_t)cpu_lduw_code(env, pc);
321 static inline uint64_t ld_code4(CPUS390XState *env, uint64_t pc)
323 return (uint64_t)(uint32_t)cpu_ldl_code(env, pc);
326 static inline uint64_t ld_code6(CPUS390XState *env, uint64_t pc)
328 return (ld_code2(env, pc) << 32) | ld_code4(env, pc + 2);
331 static inline int get_mem_index(DisasContext *s)
333 switch (s->tb->flags & FLAG_MASK_ASC) {
334 case PSW_ASC_PRIMARY >> 32:
336 case PSW_ASC_SECONDARY >> 32:
338 case PSW_ASC_HOME >> 32:
346 static void gen_exception(int excp)
348 TCGv_i32 tmp = tcg_const_i32(excp);
349 gen_helper_exception(cpu_env, tmp);
350 tcg_temp_free_i32(tmp);
353 static void gen_program_exception(DisasContext *s, int code)
357 /* Remember what pgm exeption this was. */
358 tmp = tcg_const_i32(code);
359 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_code));
360 tcg_temp_free_i32(tmp);
362 tmp = tcg_const_i32(s->next_pc - s->pc);
363 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_ilen));
364 tcg_temp_free_i32(tmp);
366 /* Advance past instruction. */
373 /* Trigger exception. */
374 gen_exception(EXCP_PGM);
377 s->is_jmp = DISAS_EXCP;
380 static inline void gen_illegal_opcode(DisasContext *s)
382 gen_program_exception(s, PGM_SPECIFICATION);
385 static inline void check_privileged(DisasContext *s)
387 if (s->tb->flags & (PSW_MASK_PSTATE >> 32)) {
388 gen_program_exception(s, PGM_PRIVILEGED);
392 static TCGv_i64 get_address(DisasContext *s, int x2, int b2, int d2)
396 /* 31-bitify the immediate part; register contents are dealt with below */
397 if (!(s->tb->flags & FLAG_MASK_64)) {
403 tmp = tcg_const_i64(d2);
404 tcg_gen_add_i64(tmp, tmp, regs[x2]);
409 tcg_gen_add_i64(tmp, tmp, regs[b2]);
413 tmp = tcg_const_i64(d2);
414 tcg_gen_add_i64(tmp, tmp, regs[b2]);
419 tmp = tcg_const_i64(d2);
422 /* 31-bit mode mask if there are values loaded from registers */
423 if (!(s->tb->flags & FLAG_MASK_64) && (x2 || b2)) {
424 tcg_gen_andi_i64(tmp, tmp, 0x7fffffffUL);
430 static inline void gen_op_movi_cc(DisasContext *s, uint32_t val)
432 s->cc_op = CC_OP_CONST0 + val;
435 static void gen_op_update1_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 dst)
437 tcg_gen_discard_i64(cc_src);
438 tcg_gen_mov_i64(cc_dst, dst);
439 tcg_gen_discard_i64(cc_vr);
443 static void gen_op_update1_cc_i32(DisasContext *s, enum cc_op op, TCGv_i32 dst)
445 tcg_gen_discard_i64(cc_src);
446 tcg_gen_extu_i32_i64(cc_dst, dst);
447 tcg_gen_discard_i64(cc_vr);
451 static void gen_op_update2_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
454 tcg_gen_mov_i64(cc_src, src);
455 tcg_gen_mov_i64(cc_dst, dst);
456 tcg_gen_discard_i64(cc_vr);
460 static void gen_op_update2_cc_i32(DisasContext *s, enum cc_op op, TCGv_i32 src,
463 tcg_gen_extu_i32_i64(cc_src, src);
464 tcg_gen_extu_i32_i64(cc_dst, dst);
465 tcg_gen_discard_i64(cc_vr);
469 static void gen_op_update3_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
470 TCGv_i64 dst, TCGv_i64 vr)
472 tcg_gen_mov_i64(cc_src, src);
473 tcg_gen_mov_i64(cc_dst, dst);
474 tcg_gen_mov_i64(cc_vr, vr);
478 static inline void set_cc_nz_u32(DisasContext *s, TCGv_i32 val)
480 gen_op_update1_cc_i32(s, CC_OP_NZ, val);
483 static inline void set_cc_nz_u64(DisasContext *s, TCGv_i64 val)
485 gen_op_update1_cc_i64(s, CC_OP_NZ, val);
488 static inline void gen_set_cc_nz_f32(DisasContext *s, TCGv_i64 val)
490 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, val);
493 static inline void gen_set_cc_nz_f64(DisasContext *s, TCGv_i64 val)
495 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, val);
498 static inline void gen_set_cc_nz_f128(DisasContext *s, TCGv_i64 vh, TCGv_i64 vl)
500 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, vh, vl);
503 static inline void cmp_32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2,
506 gen_op_update2_cc_i32(s, cond, v1, v2);
509 static inline void cmp_64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2,
512 gen_op_update2_cc_i64(s, cond, v1, v2);
515 static inline void cmp_s32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2)
517 cmp_32(s, v1, v2, CC_OP_LTGT_32);
520 static inline void cmp_u32(DisasContext *s, TCGv_i32 v1, TCGv_i32 v2)
522 cmp_32(s, v1, v2, CC_OP_LTUGTU_32);
525 static inline void cmp_s32c(DisasContext *s, TCGv_i32 v1, int32_t v2)
527 /* XXX optimize for the constant? put it in s? */
528 TCGv_i32 tmp = tcg_const_i32(v2);
529 cmp_32(s, v1, tmp, CC_OP_LTGT_32);
530 tcg_temp_free_i32(tmp);
533 static inline void cmp_u32c(DisasContext *s, TCGv_i32 v1, uint32_t v2)
535 TCGv_i32 tmp = tcg_const_i32(v2);
536 cmp_32(s, v1, tmp, CC_OP_LTUGTU_32);
537 tcg_temp_free_i32(tmp);
540 static inline void cmp_s64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2)
542 cmp_64(s, v1, v2, CC_OP_LTGT_64);
545 static inline void cmp_u64(DisasContext *s, TCGv_i64 v1, TCGv_i64 v2)
547 cmp_64(s, v1, v2, CC_OP_LTUGTU_64);
550 static inline void cmp_s64c(DisasContext *s, TCGv_i64 v1, int64_t v2)
552 TCGv_i64 tmp = tcg_const_i64(v2);
554 tcg_temp_free_i64(tmp);
557 static inline void cmp_u64c(DisasContext *s, TCGv_i64 v1, uint64_t v2)
559 TCGv_i64 tmp = tcg_const_i64(v2);
561 tcg_temp_free_i64(tmp);
564 static inline void set_cc_s32(DisasContext *s, TCGv_i32 val)
566 gen_op_update1_cc_i32(s, CC_OP_LTGT0_32, val);
569 static inline void set_cc_s64(DisasContext *s, TCGv_i64 val)
571 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, val);
574 /* CC value is in env->cc_op */
575 static inline void set_cc_static(DisasContext *s)
577 tcg_gen_discard_i64(cc_src);
578 tcg_gen_discard_i64(cc_dst);
579 tcg_gen_discard_i64(cc_vr);
580 s->cc_op = CC_OP_STATIC;
583 static inline void gen_op_set_cc_op(DisasContext *s)
585 if (s->cc_op != CC_OP_DYNAMIC && s->cc_op != CC_OP_STATIC) {
586 tcg_gen_movi_i32(cc_op, s->cc_op);
590 static inline void gen_update_cc_op(DisasContext *s)
595 /* calculates cc into cc_op */
596 static void gen_op_calc_cc(DisasContext *s)
598 TCGv_i32 local_cc_op = tcg_const_i32(s->cc_op);
599 TCGv_i64 dummy = tcg_const_i64(0);
606 /* s->cc_op is the cc value */
607 tcg_gen_movi_i32(cc_op, s->cc_op - CC_OP_CONST0);
610 /* env->cc_op already is the cc value */
625 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, dummy, cc_dst, dummy);
630 case CC_OP_LTUGTU_32:
631 case CC_OP_LTUGTU_64:
638 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, dummy);
653 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, cc_vr);
656 /* unknown operation - assume 3 arguments and cc_op in env */
657 gen_helper_calc_cc(cc_op, cpu_env, cc_op, cc_src, cc_dst, cc_vr);
663 tcg_temp_free_i32(local_cc_op);
664 tcg_temp_free_i64(dummy);
666 /* We now have cc in cc_op as constant */
670 static inline void decode_rr(DisasContext *s, uint64_t insn, int *r1, int *r2)
674 *r1 = (insn >> 4) & 0xf;
678 static inline TCGv_i64 decode_rx(DisasContext *s, uint64_t insn, int *r1,
679 int *x2, int *b2, int *d2)
683 *r1 = (insn >> 20) & 0xf;
684 *x2 = (insn >> 16) & 0xf;
685 *b2 = (insn >> 12) & 0xf;
688 return get_address(s, *x2, *b2, *d2);
691 static inline void decode_rs(DisasContext *s, uint64_t insn, int *r1, int *r3,
696 *r1 = (insn >> 20) & 0xf;
698 *r3 = (insn >> 16) & 0xf;
699 *b2 = (insn >> 12) & 0xf;
703 static inline TCGv_i64 decode_si(DisasContext *s, uint64_t insn, int *i2,
708 *i2 = (insn >> 16) & 0xff;
709 *b1 = (insn >> 12) & 0xf;
712 return get_address(s, 0, *b1, *d1);
715 static int use_goto_tb(DisasContext *s, uint64_t dest)
717 /* NOTE: we handle the case where the TB spans two pages here */
718 return (((dest & TARGET_PAGE_MASK) == (s->tb->pc & TARGET_PAGE_MASK)
719 || (dest & TARGET_PAGE_MASK) == ((s->pc - 1) & TARGET_PAGE_MASK))
720 && !s->singlestep_enabled
721 && !(s->tb->cflags & CF_LAST_IO));
724 static inline void gen_goto_tb(DisasContext *s, int tb_num, target_ulong pc)
728 if (use_goto_tb(s, pc)) {
729 tcg_gen_goto_tb(tb_num);
730 tcg_gen_movi_i64(psw_addr, pc);
731 tcg_gen_exit_tb((tcg_target_long)s->tb + tb_num);
733 /* jump to another page: currently not optimized */
734 tcg_gen_movi_i64(psw_addr, pc);
739 static inline void account_noninline_branch(DisasContext *s, int cc_op)
741 #ifdef DEBUG_INLINE_BRANCHES
742 inline_branch_miss[cc_op]++;
746 static inline void account_inline_branch(DisasContext *s, int cc_op)
748 #ifdef DEBUG_INLINE_BRANCHES
749 inline_branch_hit[cc_op]++;
753 /* Table of mask values to comparison codes, given a comparison as input.
754 For a true comparison CC=3 will never be set, but we treat this
755 conservatively for possible use when CC=3 indicates overflow. */
756 static const TCGCond ltgt_cond[16] = {
757 TCG_COND_NEVER, TCG_COND_NEVER, /* | | | x */
758 TCG_COND_GT, TCG_COND_NEVER, /* | | GT | x */
759 TCG_COND_LT, TCG_COND_NEVER, /* | LT | | x */
760 TCG_COND_NE, TCG_COND_NEVER, /* | LT | GT | x */
761 TCG_COND_EQ, TCG_COND_NEVER, /* EQ | | | x */
762 TCG_COND_GE, TCG_COND_NEVER, /* EQ | | GT | x */
763 TCG_COND_LE, TCG_COND_NEVER, /* EQ | LT | | x */
764 TCG_COND_ALWAYS, TCG_COND_ALWAYS, /* EQ | LT | GT | x */
767 /* Table of mask values to comparison codes, given a logic op as input.
768 For such, only CC=0 and CC=1 should be possible. */
769 static const TCGCond nz_cond[16] = {
771 TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER,
773 TCG_COND_NE, TCG_COND_NE, TCG_COND_NE, TCG_COND_NE,
775 TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ,
776 /* EQ | NE | x | x */
777 TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS,
780 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
781 details required to generate a TCG comparison. */
782 static void disas_jcc(DisasContext *s, DisasCompare *c, uint32_t mask)
785 enum cc_op old_cc_op = s->cc_op;
787 if (mask == 15 || mask == 0) {
788 c->cond = (mask ? TCG_COND_ALWAYS : TCG_COND_NEVER);
791 c->g1 = c->g2 = true;
796 /* Find the TCG condition for the mask + cc op. */
802 cond = ltgt_cond[mask];
803 if (cond == TCG_COND_NEVER) {
806 account_inline_branch(s, old_cc_op);
809 case CC_OP_LTUGTU_32:
810 case CC_OP_LTUGTU_64:
811 cond = tcg_unsigned_cond(ltgt_cond[mask]);
812 if (cond == TCG_COND_NEVER) {
815 account_inline_branch(s, old_cc_op);
819 cond = nz_cond[mask];
820 if (cond == TCG_COND_NEVER) {
823 account_inline_branch(s, old_cc_op);
838 account_inline_branch(s, old_cc_op);
853 account_inline_branch(s, old_cc_op);
857 switch (mask & 0xa) {
858 case 8: /* src == 0 -> no one bit found */
861 case 2: /* src != 0 -> one bit found */
867 account_inline_branch(s, old_cc_op);
872 /* Calculate cc value. */
877 /* Jump based on CC. We'll load up the real cond below;
878 the assignment here merely avoids a compiler warning. */
879 account_noninline_branch(s, old_cc_op);
880 old_cc_op = CC_OP_STATIC;
881 cond = TCG_COND_NEVER;
885 /* Load up the arguments of the comparison. */
887 c->g1 = c->g2 = false;
891 c->u.s32.a = tcg_temp_new_i32();
892 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_dst);
893 c->u.s32.b = tcg_const_i32(0);
896 case CC_OP_LTUGTU_32:
898 c->u.s32.a = tcg_temp_new_i32();
899 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_src);
900 c->u.s32.b = tcg_temp_new_i32();
901 tcg_gen_trunc_i64_i32(c->u.s32.b, cc_dst);
908 c->u.s64.b = tcg_const_i64(0);
912 case CC_OP_LTUGTU_64:
915 c->g1 = c->g2 = true;
921 c->u.s64.a = tcg_temp_new_i64();
922 c->u.s64.b = tcg_const_i64(0);
923 tcg_gen_and_i64(c->u.s64.a, cc_src, cc_dst);
931 case 0x8 | 0x4 | 0x2: /* cc != 3 */
933 c->u.s32.b = tcg_const_i32(3);
935 case 0x8 | 0x4 | 0x1: /* cc != 2 */
937 c->u.s32.b = tcg_const_i32(2);
939 case 0x8 | 0x2 | 0x1: /* cc != 1 */
941 c->u.s32.b = tcg_const_i32(1);
943 case 0x8 | 0x2: /* cc == 0 ||Â cc == 2 => (cc & 1) == 0 */
946 c->u.s32.a = tcg_temp_new_i32();
947 c->u.s32.b = tcg_const_i32(0);
948 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
950 case 0x8 | 0x4: /* cc < 2 */
952 c->u.s32.b = tcg_const_i32(2);
954 case 0x8: /* cc == 0 */
956 c->u.s32.b = tcg_const_i32(0);
958 case 0x4 | 0x2 | 0x1: /* cc != 0 */
960 c->u.s32.b = tcg_const_i32(0);
962 case 0x4 | 0x1: /* cc == 1 ||Â cc == 3 => (cc & 1) != 0 */
965 c->u.s32.a = tcg_temp_new_i32();
966 c->u.s32.b = tcg_const_i32(0);
967 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
969 case 0x4: /* cc == 1 */
971 c->u.s32.b = tcg_const_i32(1);
973 case 0x2 | 0x1: /* cc > 1 */
975 c->u.s32.b = tcg_const_i32(1);
977 case 0x2: /* cc == 2 */
979 c->u.s32.b = tcg_const_i32(2);
981 case 0x1: /* cc == 3 */
983 c->u.s32.b = tcg_const_i32(3);
986 /* CC is masked by something else: (8 >> cc) & mask. */
989 c->u.s32.a = tcg_const_i32(8);
990 c->u.s32.b = tcg_const_i32(0);
991 tcg_gen_shr_i32(c->u.s32.a, c->u.s32.a, cc_op);
992 tcg_gen_andi_i32(c->u.s32.a, c->u.s32.a, mask);
1003 static void free_compare(DisasCompare *c)
1007 tcg_temp_free_i64(c->u.s64.a);
1009 tcg_temp_free_i32(c->u.s32.a);
1014 tcg_temp_free_i64(c->u.s64.b);
1016 tcg_temp_free_i32(c->u.s32.b);
1021 static void disas_b2(CPUS390XState *env, DisasContext *s, int op,
1024 #ifndef CONFIG_USER_ONLY
1025 TCGv_i64 tmp, tmp2, tmp3;
1030 r1 = (insn >> 4) & 0xf;
1033 LOG_DISAS("disas_b2: op 0x%x r1 %d r2 %d\n", op, r1, r2);
1036 case 0xb1: /* STFL D2(B2) [S] */
1037 /* Store Facility List (CPU features) at 200 */
1038 check_privileged(s);
1039 tmp2 = tcg_const_i64(0xc0000000);
1040 tmp = tcg_const_i64(200);
1041 tcg_gen_qemu_st32(tmp2, tmp, get_mem_index(s));
1042 tcg_temp_free_i64(tmp2);
1043 tcg_temp_free_i64(tmp);
1045 case 0xb2: /* LPSWE D2(B2) [S] */
1046 /* Load PSW Extended */
1047 check_privileged(s);
1048 decode_rs(s, insn, &r1, &r3, &b2, &d2);
1049 tmp = get_address(s, 0, b2, d2);
1050 tmp2 = tcg_temp_new_i64();
1051 tmp3 = tcg_temp_new_i64();
1052 tcg_gen_qemu_ld64(tmp2, tmp, get_mem_index(s));
1053 tcg_gen_addi_i64(tmp, tmp, 8);
1054 tcg_gen_qemu_ld64(tmp3, tmp, get_mem_index(s));
1055 gen_helper_load_psw(cpu_env, tmp2, tmp3);
1056 /* we need to keep cc_op intact */
1057 s->is_jmp = DISAS_JUMP;
1058 tcg_temp_free_i64(tmp);
1059 tcg_temp_free_i64(tmp2);
1060 tcg_temp_free_i64(tmp3);
1062 case 0x20: /* SERVC R1,R2 [RRE] */
1063 /* SCLP Service call (PV hypercall) */
1064 check_privileged(s);
1065 potential_page_fault(s);
1066 tmp32_1 = load_reg32(r2);
1068 gen_helper_servc(cc_op, cpu_env, tmp32_1, tmp);
1070 tcg_temp_free_i32(tmp32_1);
1071 tcg_temp_free_i64(tmp);
1075 LOG_DISAS("illegal b2 operation 0x%x\n", op);
1076 gen_illegal_opcode(s);
1077 #ifndef CONFIG_USER_ONLY
1083 static void disas_s390_insn(CPUS390XState *env, DisasContext *s)
1089 opc = cpu_ldub_code(env, s->pc);
1090 LOG_DISAS("opc 0x%x\n", opc);
1094 insn = ld_code4(env, s->pc);
1095 op = (insn >> 16) & 0xff;
1096 disas_b2(env, s, op, insn);
1099 qemu_log_mask(LOG_UNIMP, "unimplemented opcode 0x%x\n", opc);
1100 gen_illegal_opcode(s);
1105 /* ====================================================================== */
1106 /* Define the insn format enumeration. */
1107 #define F0(N) FMT_##N,
1108 #define F1(N, X1) F0(N)
1109 #define F2(N, X1, X2) F0(N)
1110 #define F3(N, X1, X2, X3) F0(N)
1111 #define F4(N, X1, X2, X3, X4) F0(N)
1112 #define F5(N, X1, X2, X3, X4, X5) F0(N)
1115 #include "insn-format.def"
1125 /* Define a structure to hold the decoded fields. We'll store each inside
1126 an array indexed by an enum. In order to conserve memory, we'll arrange
1127 for fields that do not exist at the same time to overlap, thus the "C"
1128 for compact. For checking purposes there is an "O" for original index
1129 as well that will be applied to availability bitmaps. */
1131 enum DisasFieldIndexO {
1154 enum DisasFieldIndexC {
1185 struct DisasFields {
1188 unsigned presentC:16;
1189 unsigned int presentO;
1193 /* This is the way fields are to be accessed out of DisasFields. */
1194 #define have_field(S, F) have_field1((S), FLD_O_##F)
1195 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
1197 static bool have_field1(const DisasFields *f, enum DisasFieldIndexO c)
1199 return (f->presentO >> c) & 1;
1202 static int get_field1(const DisasFields *f, enum DisasFieldIndexO o,
1203 enum DisasFieldIndexC c)
1205 assert(have_field1(f, o));
1209 /* Describe the layout of each field in each format. */
1210 typedef struct DisasField {
1212 unsigned int size:8;
1213 unsigned int type:2;
1214 unsigned int indexC:6;
1215 enum DisasFieldIndexO indexO:8;
1218 typedef struct DisasFormatInfo {
1219 DisasField op[NUM_C_FIELD];
1222 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
1223 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
1224 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1225 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
1226 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1227 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1228 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
1229 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1230 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1231 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1232 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1233 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1234 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
1235 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
1237 #define F0(N) { { } },
1238 #define F1(N, X1) { { X1 } },
1239 #define F2(N, X1, X2) { { X1, X2 } },
1240 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
1241 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
1242 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
1244 static const DisasFormatInfo format_info[] = {
1245 #include "insn-format.def"
1263 /* Generally, we'll extract operands into this structures, operate upon
1264 them, and store them back. See the "in1", "in2", "prep", "wout" sets
1265 of routines below for more details. */
1267 bool g_out, g_out2, g_in1, g_in2;
1268 TCGv_i64 out, out2, in1, in2;
1272 /* Return values from translate_one, indicating the state of the TB. */
1274 /* Continue the TB. */
1276 /* We have emitted one or more goto_tb. No fixup required. */
1278 /* We are not using a goto_tb (for whatever reason), but have updated
1279 the PC (for whatever reason), so there's no need to do it again on
1282 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1283 updated the PC for the next instruction to be executed. */
1285 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1286 No following code will be executed. */
1290 typedef enum DisasFacility {
1291 FAC_Z, /* zarch (default) */
1292 FAC_CASS, /* compare and swap and store */
1293 FAC_CASS2, /* compare and swap and store 2*/
1294 FAC_DFP, /* decimal floating point */
1295 FAC_DFPR, /* decimal floating point rounding */
1296 FAC_DO, /* distinct operands */
1297 FAC_EE, /* execute extensions */
1298 FAC_EI, /* extended immediate */
1299 FAC_FPE, /* floating point extension */
1300 FAC_FPSSH, /* floating point support sign handling */
1301 FAC_FPRGR, /* FPR-GR transfer */
1302 FAC_GIE, /* general instructions extension */
1303 FAC_HFP_MA, /* HFP multiply-and-add/subtract */
1304 FAC_HW, /* high-word */
1305 FAC_IEEEE_SIM, /* IEEE exception sumilation */
1306 FAC_LOC, /* load/store on condition */
1307 FAC_LD, /* long displacement */
1308 FAC_PC, /* population count */
1309 FAC_SCF, /* store clock fast */
1310 FAC_SFLE, /* store facility list extended */
1316 DisasFacility fac:6;
1320 void (*help_in1)(DisasContext *, DisasFields *, DisasOps *);
1321 void (*help_in2)(DisasContext *, DisasFields *, DisasOps *);
1322 void (*help_prep)(DisasContext *, DisasFields *, DisasOps *);
1323 void (*help_wout)(DisasContext *, DisasFields *, DisasOps *);
1324 void (*help_cout)(DisasContext *, DisasOps *);
1325 ExitStatus (*help_op)(DisasContext *, DisasOps *);
1330 /* ====================================================================== */
1331 /* Miscelaneous helpers, used by several operations. */
1333 static void help_l2_shift(DisasContext *s, DisasFields *f,
1334 DisasOps *o, int mask)
1336 int b2 = get_field(f, b2);
1337 int d2 = get_field(f, d2);
1340 o->in2 = tcg_const_i64(d2 & mask);
1342 o->in2 = get_address(s, 0, b2, d2);
1343 tcg_gen_andi_i64(o->in2, o->in2, mask);
1347 static ExitStatus help_goto_direct(DisasContext *s, uint64_t dest)
1349 if (dest == s->next_pc) {
1352 if (use_goto_tb(s, dest)) {
1353 gen_update_cc_op(s);
1355 tcg_gen_movi_i64(psw_addr, dest);
1356 tcg_gen_exit_tb((tcg_target_long)s->tb);
1357 return EXIT_GOTO_TB;
1359 tcg_gen_movi_i64(psw_addr, dest);
1360 return EXIT_PC_UPDATED;
1364 static ExitStatus help_branch(DisasContext *s, DisasCompare *c,
1365 bool is_imm, int imm, TCGv_i64 cdest)
1368 uint64_t dest = s->pc + 2 * imm;
1371 /* Take care of the special cases first. */
1372 if (c->cond == TCG_COND_NEVER) {
1377 if (dest == s->next_pc) {
1378 /* Branch to next. */
1382 if (c->cond == TCG_COND_ALWAYS) {
1383 ret = help_goto_direct(s, dest);
1387 if (TCGV_IS_UNUSED_I64(cdest)) {
1388 /* E.g. bcr %r0 -> no branch. */
1392 if (c->cond == TCG_COND_ALWAYS) {
1393 tcg_gen_mov_i64(psw_addr, cdest);
1394 ret = EXIT_PC_UPDATED;
1399 if (use_goto_tb(s, s->next_pc)) {
1400 if (is_imm && use_goto_tb(s, dest)) {
1401 /* Both exits can use goto_tb. */
1402 gen_update_cc_op(s);
1404 lab = gen_new_label();
1406 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1408 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1411 /* Branch not taken. */
1413 tcg_gen_movi_i64(psw_addr, s->next_pc);
1414 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1419 tcg_gen_movi_i64(psw_addr, dest);
1420 tcg_gen_exit_tb((tcg_target_long)s->tb + 1);
1424 /* Fallthru can use goto_tb, but taken branch cannot. */
1425 /* Store taken branch destination before the brcond. This
1426 avoids having to allocate a new local temp to hold it.
1427 We'll overwrite this in the not taken case anyway. */
1429 tcg_gen_mov_i64(psw_addr, cdest);
1432 lab = gen_new_label();
1434 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1436 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1439 /* Branch not taken. */
1440 gen_update_cc_op(s);
1442 tcg_gen_movi_i64(psw_addr, s->next_pc);
1443 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1447 tcg_gen_movi_i64(psw_addr, dest);
1449 ret = EXIT_PC_UPDATED;
1452 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1453 Most commonly we're single-stepping or some other condition that
1454 disables all use of goto_tb. Just update the PC and exit. */
1456 TCGv_i64 next = tcg_const_i64(s->next_pc);
1458 cdest = tcg_const_i64(dest);
1462 tcg_gen_movcond_i64(c->cond, psw_addr, c->u.s64.a, c->u.s64.b,
1465 TCGv_i32 t0 = tcg_temp_new_i32();
1466 TCGv_i64 t1 = tcg_temp_new_i64();
1467 TCGv_i64 z = tcg_const_i64(0);
1468 tcg_gen_setcond_i32(c->cond, t0, c->u.s32.a, c->u.s32.b);
1469 tcg_gen_extu_i32_i64(t1, t0);
1470 tcg_temp_free_i32(t0);
1471 tcg_gen_movcond_i64(TCG_COND_NE, psw_addr, t1, z, cdest, next);
1472 tcg_temp_free_i64(t1);
1473 tcg_temp_free_i64(z);
1477 tcg_temp_free_i64(cdest);
1479 tcg_temp_free_i64(next);
1481 ret = EXIT_PC_UPDATED;
1489 /* ====================================================================== */
1490 /* The operations. These perform the bulk of the work for any insn,
1491 usually after the operands have been loaded and output initialized. */
1493 static ExitStatus op_abs(DisasContext *s, DisasOps *o)
1495 gen_helper_abs_i64(o->out, o->in2);
1499 static ExitStatus op_absf32(DisasContext *s, DisasOps *o)
1501 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffull);
1505 static ExitStatus op_absf64(DisasContext *s, DisasOps *o)
1507 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffffffffffull);
1511 static ExitStatus op_absf128(DisasContext *s, DisasOps *o)
1513 tcg_gen_andi_i64(o->out, o->in1, 0x7fffffffffffffffull);
1514 tcg_gen_mov_i64(o->out2, o->in2);
1518 static ExitStatus op_add(DisasContext *s, DisasOps *o)
1520 tcg_gen_add_i64(o->out, o->in1, o->in2);
1524 static ExitStatus op_addc(DisasContext *s, DisasOps *o)
1528 tcg_gen_add_i64(o->out, o->in1, o->in2);
1530 /* XXX possible optimization point */
1532 cc = tcg_temp_new_i64();
1533 tcg_gen_extu_i32_i64(cc, cc_op);
1534 tcg_gen_shri_i64(cc, cc, 1);
1536 tcg_gen_add_i64(o->out, o->out, cc);
1537 tcg_temp_free_i64(cc);
1541 static ExitStatus op_aeb(DisasContext *s, DisasOps *o)
1543 gen_helper_aeb(o->out, cpu_env, o->in1, o->in2);
1547 static ExitStatus op_adb(DisasContext *s, DisasOps *o)
1549 gen_helper_adb(o->out, cpu_env, o->in1, o->in2);
1553 static ExitStatus op_axb(DisasContext *s, DisasOps *o)
1555 gen_helper_axb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1556 return_low128(o->out2);
1560 static ExitStatus op_and(DisasContext *s, DisasOps *o)
1562 tcg_gen_and_i64(o->out, o->in1, o->in2);
1566 static ExitStatus op_andi(DisasContext *s, DisasOps *o)
1568 int shift = s->insn->data & 0xff;
1569 int size = s->insn->data >> 8;
1570 uint64_t mask = ((1ull << size) - 1) << shift;
1573 tcg_gen_shli_i64(o->in2, o->in2, shift);
1574 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
1575 tcg_gen_and_i64(o->out, o->in1, o->in2);
1577 /* Produce the CC from only the bits manipulated. */
1578 tcg_gen_andi_i64(cc_dst, o->out, mask);
1579 set_cc_nz_u64(s, cc_dst);
1583 static ExitStatus op_bas(DisasContext *s, DisasOps *o)
1585 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1586 if (!TCGV_IS_UNUSED_I64(o->in2)) {
1587 tcg_gen_mov_i64(psw_addr, o->in2);
1588 return EXIT_PC_UPDATED;
1594 static ExitStatus op_basi(DisasContext *s, DisasOps *o)
1596 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1597 return help_goto_direct(s, s->pc + 2 * get_field(s->fields, i2));
1600 static ExitStatus op_bc(DisasContext *s, DisasOps *o)
1602 int m1 = get_field(s->fields, m1);
1603 bool is_imm = have_field(s->fields, i2);
1604 int imm = is_imm ? get_field(s->fields, i2) : 0;
1607 disas_jcc(s, &c, m1);
1608 return help_branch(s, &c, is_imm, imm, o->in2);
1611 static ExitStatus op_bct32(DisasContext *s, DisasOps *o)
1613 int r1 = get_field(s->fields, r1);
1614 bool is_imm = have_field(s->fields, i2);
1615 int imm = is_imm ? get_field(s->fields, i2) : 0;
1619 c.cond = TCG_COND_NE;
1624 t = tcg_temp_new_i64();
1625 tcg_gen_subi_i64(t, regs[r1], 1);
1626 store_reg32_i64(r1, t);
1627 c.u.s32.a = tcg_temp_new_i32();
1628 c.u.s32.b = tcg_const_i32(0);
1629 tcg_gen_trunc_i64_i32(c.u.s32.a, t);
1630 tcg_temp_free_i64(t);
1632 return help_branch(s, &c, is_imm, imm, o->in2);
1635 static ExitStatus op_bct64(DisasContext *s, DisasOps *o)
1637 int r1 = get_field(s->fields, r1);
1638 bool is_imm = have_field(s->fields, i2);
1639 int imm = is_imm ? get_field(s->fields, i2) : 0;
1642 c.cond = TCG_COND_NE;
1647 tcg_gen_subi_i64(regs[r1], regs[r1], 1);
1648 c.u.s64.a = regs[r1];
1649 c.u.s64.b = tcg_const_i64(0);
1651 return help_branch(s, &c, is_imm, imm, o->in2);
1654 static ExitStatus op_ceb(DisasContext *s, DisasOps *o)
1656 gen_helper_ceb(cc_op, cpu_env, o->in1, o->in2);
1661 static ExitStatus op_cdb(DisasContext *s, DisasOps *o)
1663 gen_helper_cdb(cc_op, cpu_env, o->in1, o->in2);
1668 static ExitStatus op_cxb(DisasContext *s, DisasOps *o)
1670 gen_helper_cxb(cc_op, cpu_env, o->out, o->out2, o->in1, o->in2);
1675 static ExitStatus op_cfeb(DisasContext *s, DisasOps *o)
1677 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1678 gen_helper_cfeb(o->out, cpu_env, o->in2, m3);
1679 tcg_temp_free_i32(m3);
1680 gen_set_cc_nz_f32(s, o->in2);
1684 static ExitStatus op_cfdb(DisasContext *s, DisasOps *o)
1686 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1687 gen_helper_cfdb(o->out, cpu_env, o->in2, m3);
1688 tcg_temp_free_i32(m3);
1689 gen_set_cc_nz_f64(s, o->in2);
1693 static ExitStatus op_cfxb(DisasContext *s, DisasOps *o)
1695 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1696 gen_helper_cfxb(o->out, cpu_env, o->in1, o->in2, m3);
1697 tcg_temp_free_i32(m3);
1698 gen_set_cc_nz_f128(s, o->in1, o->in2);
1702 static ExitStatus op_cgeb(DisasContext *s, DisasOps *o)
1704 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1705 gen_helper_cgeb(o->out, cpu_env, o->in2, m3);
1706 tcg_temp_free_i32(m3);
1707 gen_set_cc_nz_f32(s, o->in2);
1711 static ExitStatus op_cgdb(DisasContext *s, DisasOps *o)
1713 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1714 gen_helper_cgdb(o->out, cpu_env, o->in2, m3);
1715 tcg_temp_free_i32(m3);
1716 gen_set_cc_nz_f64(s, o->in2);
1720 static ExitStatus op_cgxb(DisasContext *s, DisasOps *o)
1722 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1723 gen_helper_cgxb(o->out, cpu_env, o->in1, o->in2, m3);
1724 tcg_temp_free_i32(m3);
1725 gen_set_cc_nz_f128(s, o->in1, o->in2);
1729 static ExitStatus op_cegb(DisasContext *s, DisasOps *o)
1731 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1732 gen_helper_cegb(o->out, cpu_env, o->in2, m3);
1733 tcg_temp_free_i32(m3);
1737 static ExitStatus op_cdgb(DisasContext *s, DisasOps *o)
1739 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1740 gen_helper_cdgb(o->out, cpu_env, o->in2, m3);
1741 tcg_temp_free_i32(m3);
1745 static ExitStatus op_cxgb(DisasContext *s, DisasOps *o)
1747 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1748 gen_helper_cxgb(o->out, cpu_env, o->in2, m3);
1749 tcg_temp_free_i32(m3);
1750 return_low128(o->out2);
1754 static ExitStatus op_cksm(DisasContext *s, DisasOps *o)
1756 int r2 = get_field(s->fields, r2);
1757 TCGv_i64 len = tcg_temp_new_i64();
1759 potential_page_fault(s);
1760 gen_helper_cksm(len, cpu_env, o->in1, o->in2, regs[r2 + 1]);
1762 return_low128(o->out);
1764 tcg_gen_add_i64(regs[r2], regs[r2], len);
1765 tcg_gen_sub_i64(regs[r2 + 1], regs[r2 + 1], len);
1766 tcg_temp_free_i64(len);
1771 static ExitStatus op_clc(DisasContext *s, DisasOps *o)
1773 int l = get_field(s->fields, l1);
1778 tcg_gen_qemu_ld8u(cc_src, o->addr1, get_mem_index(s));
1779 tcg_gen_qemu_ld8u(cc_dst, o->in2, get_mem_index(s));
1782 tcg_gen_qemu_ld16u(cc_src, o->addr1, get_mem_index(s));
1783 tcg_gen_qemu_ld16u(cc_dst, o->in2, get_mem_index(s));
1786 tcg_gen_qemu_ld32u(cc_src, o->addr1, get_mem_index(s));
1787 tcg_gen_qemu_ld32u(cc_dst, o->in2, get_mem_index(s));
1790 tcg_gen_qemu_ld64(cc_src, o->addr1, get_mem_index(s));
1791 tcg_gen_qemu_ld64(cc_dst, o->in2, get_mem_index(s));
1794 potential_page_fault(s);
1795 vl = tcg_const_i32(l);
1796 gen_helper_clc(cc_op, cpu_env, vl, o->addr1, o->in2);
1797 tcg_temp_free_i32(vl);
1801 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, cc_src, cc_dst);
1805 static ExitStatus op_clcle(DisasContext *s, DisasOps *o)
1807 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1808 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1809 potential_page_fault(s);
1810 gen_helper_clcle(cc_op, cpu_env, r1, o->in2, r3);
1811 tcg_temp_free_i32(r1);
1812 tcg_temp_free_i32(r3);
1817 static ExitStatus op_clm(DisasContext *s, DisasOps *o)
1819 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1820 TCGv_i32 t1 = tcg_temp_new_i32();
1821 tcg_gen_trunc_i64_i32(t1, o->in1);
1822 potential_page_fault(s);
1823 gen_helper_clm(cc_op, cpu_env, t1, m3, o->in2);
1825 tcg_temp_free_i32(t1);
1826 tcg_temp_free_i32(m3);
1830 static ExitStatus op_clst(DisasContext *s, DisasOps *o)
1832 potential_page_fault(s);
1833 gen_helper_clst(o->in1, cpu_env, regs[0], o->in1, o->in2);
1835 return_low128(o->in2);
1839 static ExitStatus op_cs(DisasContext *s, DisasOps *o)
1841 int r3 = get_field(s->fields, r3);
1842 potential_page_fault(s);
1843 gen_helper_cs(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1848 static ExitStatus op_csg(DisasContext *s, DisasOps *o)
1850 int r3 = get_field(s->fields, r3);
1851 potential_page_fault(s);
1852 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1857 #ifndef CONFIG_USER_ONLY
1858 static ExitStatus op_csp(DisasContext *s, DisasOps *o)
1860 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1861 check_privileged(s);
1862 gen_helper_csp(cc_op, cpu_env, r1, o->in2);
1863 tcg_temp_free_i32(r1);
1869 static ExitStatus op_cds(DisasContext *s, DisasOps *o)
1871 int r3 = get_field(s->fields, r3);
1872 TCGv_i64 in3 = tcg_temp_new_i64();
1873 tcg_gen_deposit_i64(in3, regs[r3 + 1], regs[r3], 32, 32);
1874 potential_page_fault(s);
1875 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, in3);
1876 tcg_temp_free_i64(in3);
1881 static ExitStatus op_cdsg(DisasContext *s, DisasOps *o)
1883 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1884 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1885 potential_page_fault(s);
1886 /* XXX rewrite in tcg */
1887 gen_helper_cdsg(cc_op, cpu_env, r1, o->in2, r3);
1892 static ExitStatus op_cvd(DisasContext *s, DisasOps *o)
1894 TCGv_i64 t1 = tcg_temp_new_i64();
1895 TCGv_i32 t2 = tcg_temp_new_i32();
1896 tcg_gen_trunc_i64_i32(t2, o->in1);
1897 gen_helper_cvd(t1, t2);
1898 tcg_temp_free_i32(t2);
1899 tcg_gen_qemu_st64(t1, o->in2, get_mem_index(s));
1900 tcg_temp_free_i64(t1);
1904 #ifndef CONFIG_USER_ONLY
1905 static ExitStatus op_diag(DisasContext *s, DisasOps *o)
1909 check_privileged(s);
1910 potential_page_fault(s);
1912 /* We pretend the format is RX_a so that D2 is the field we want. */
1913 tmp = tcg_const_i32(get_field(s->fields, d2) & 0xfff);
1914 gen_helper_diag(regs[2], cpu_env, tmp, regs[2], regs[1]);
1915 tcg_temp_free_i32(tmp);
1920 static ExitStatus op_divs32(DisasContext *s, DisasOps *o)
1922 gen_helper_divs32(o->out2, cpu_env, o->in1, o->in2);
1923 return_low128(o->out);
1927 static ExitStatus op_divu32(DisasContext *s, DisasOps *o)
1929 gen_helper_divu32(o->out2, cpu_env, o->in1, o->in2);
1930 return_low128(o->out);
1934 static ExitStatus op_divs64(DisasContext *s, DisasOps *o)
1936 gen_helper_divs64(o->out2, cpu_env, o->in1, o->in2);
1937 return_low128(o->out);
1941 static ExitStatus op_divu64(DisasContext *s, DisasOps *o)
1943 gen_helper_divu64(o->out2, cpu_env, o->out, o->out2, o->in2);
1944 return_low128(o->out);
1948 static ExitStatus op_deb(DisasContext *s, DisasOps *o)
1950 gen_helper_deb(o->out, cpu_env, o->in1, o->in2);
1954 static ExitStatus op_ddb(DisasContext *s, DisasOps *o)
1956 gen_helper_ddb(o->out, cpu_env, o->in1, o->in2);
1960 static ExitStatus op_dxb(DisasContext *s, DisasOps *o)
1962 gen_helper_dxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1963 return_low128(o->out2);
1967 static ExitStatus op_ear(DisasContext *s, DisasOps *o)
1969 int r2 = get_field(s->fields, r2);
1970 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, aregs[r2]));
1974 static ExitStatus op_efpc(DisasContext *s, DisasOps *o)
1976 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, fpc));
1980 static ExitStatus op_ex(DisasContext *s, DisasOps *o)
1982 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
1983 tb->flags, (ab)use the tb->cs_base field as the address of
1984 the template in memory, and grab 8 bits of tb->flags/cflags for
1985 the contents of the register. We would then recognize all this
1986 in gen_intermediate_code_internal, generating code for exactly
1987 one instruction. This new TB then gets executed normally.
1989 On the other hand, this seems to be mostly used for modifying
1990 MVC inside of memcpy, which needs a helper call anyway. So
1991 perhaps this doesn't bear thinking about any further. */
1998 tmp = tcg_const_i64(s->next_pc);
1999 gen_helper_ex(cc_op, cpu_env, cc_op, o->in1, o->in2, tmp);
2000 tcg_temp_free_i64(tmp);
2006 static ExitStatus op_flogr(DisasContext *s, DisasOps *o)
2008 /* We'll use the original input for cc computation, since we get to
2009 compare that against 0, which ought to be better than comparing
2010 the real output against 64. It also lets cc_dst be a convenient
2011 temporary during our computation. */
2012 gen_op_update1_cc_i64(s, CC_OP_FLOGR, o->in2);
2014 /* R1 = IN ? CLZ(IN) : 64. */
2015 gen_helper_clz(o->out, o->in2);
2017 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
2018 value by 64, which is undefined. But since the shift is 64 iff the
2019 input is zero, we still get the correct result after and'ing. */
2020 tcg_gen_movi_i64(o->out2, 0x8000000000000000ull);
2021 tcg_gen_shr_i64(o->out2, o->out2, o->out);
2022 tcg_gen_andc_i64(o->out2, cc_dst, o->out2);
2026 static ExitStatus op_icm(DisasContext *s, DisasOps *o)
2028 int m3 = get_field(s->fields, m3);
2029 int pos, len, base = s->insn->data;
2030 TCGv_i64 tmp = tcg_temp_new_i64();
2035 /* Effectively a 32-bit load. */
2036 tcg_gen_qemu_ld32u(tmp, o->in2, get_mem_index(s));
2043 /* Effectively a 16-bit load. */
2044 tcg_gen_qemu_ld16u(tmp, o->in2, get_mem_index(s));
2052 /* Effectively an 8-bit load. */
2053 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2058 pos = base + ctz32(m3) * 8;
2059 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, len);
2060 ccm = ((1ull << len) - 1) << pos;
2064 /* This is going to be a sequence of loads and inserts. */
2065 pos = base + 32 - 8;
2069 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2070 tcg_gen_addi_i64(o->in2, o->in2, 1);
2071 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, 8);
2074 m3 = (m3 << 1) & 0xf;
2080 tcg_gen_movi_i64(tmp, ccm);
2081 gen_op_update2_cc_i64(s, CC_OP_ICM, tmp, o->out);
2082 tcg_temp_free_i64(tmp);
2086 static ExitStatus op_insi(DisasContext *s, DisasOps *o)
2088 int shift = s->insn->data & 0xff;
2089 int size = s->insn->data >> 8;
2090 tcg_gen_deposit_i64(o->out, o->in1, o->in2, shift, size);
2094 static ExitStatus op_ipm(DisasContext *s, DisasOps *o)
2099 tcg_gen_andi_i64(o->out, o->out, ~0xff000000ull);
2101 t1 = tcg_temp_new_i64();
2102 tcg_gen_shli_i64(t1, psw_mask, 20);
2103 tcg_gen_shri_i64(t1, t1, 36);
2104 tcg_gen_or_i64(o->out, o->out, t1);
2106 tcg_gen_extu_i32_i64(t1, cc_op);
2107 tcg_gen_shli_i64(t1, t1, 28);
2108 tcg_gen_or_i64(o->out, o->out, t1);
2109 tcg_temp_free_i64(t1);
2113 #ifndef CONFIG_USER_ONLY
2114 static ExitStatus op_ipte(DisasContext *s, DisasOps *o)
2116 check_privileged(s);
2117 gen_helper_ipte(cpu_env, o->in1, o->in2);
2121 static ExitStatus op_iske(DisasContext *s, DisasOps *o)
2123 check_privileged(s);
2124 gen_helper_iske(o->out, cpu_env, o->in2);
2129 static ExitStatus op_ldeb(DisasContext *s, DisasOps *o)
2131 gen_helper_ldeb(o->out, cpu_env, o->in2);
2135 static ExitStatus op_ledb(DisasContext *s, DisasOps *o)
2137 gen_helper_ledb(o->out, cpu_env, o->in2);
2141 static ExitStatus op_ldxb(DisasContext *s, DisasOps *o)
2143 gen_helper_ldxb(o->out, cpu_env, o->in1, o->in2);
2147 static ExitStatus op_lexb(DisasContext *s, DisasOps *o)
2149 gen_helper_lexb(o->out, cpu_env, o->in1, o->in2);
2153 static ExitStatus op_lxdb(DisasContext *s, DisasOps *o)
2155 gen_helper_lxdb(o->out, cpu_env, o->in2);
2156 return_low128(o->out2);
2160 static ExitStatus op_lxeb(DisasContext *s, DisasOps *o)
2162 gen_helper_lxeb(o->out, cpu_env, o->in2);
2163 return_low128(o->out2);
2167 static ExitStatus op_llgt(DisasContext *s, DisasOps *o)
2169 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffff);
2173 static ExitStatus op_ld8s(DisasContext *s, DisasOps *o)
2175 tcg_gen_qemu_ld8s(o->out, o->in2, get_mem_index(s));
2179 static ExitStatus op_ld8u(DisasContext *s, DisasOps *o)
2181 tcg_gen_qemu_ld8u(o->out, o->in2, get_mem_index(s));
2185 static ExitStatus op_ld16s(DisasContext *s, DisasOps *o)
2187 tcg_gen_qemu_ld16s(o->out, o->in2, get_mem_index(s));
2191 static ExitStatus op_ld16u(DisasContext *s, DisasOps *o)
2193 tcg_gen_qemu_ld16u(o->out, o->in2, get_mem_index(s));
2197 static ExitStatus op_ld32s(DisasContext *s, DisasOps *o)
2199 tcg_gen_qemu_ld32s(o->out, o->in2, get_mem_index(s));
2203 static ExitStatus op_ld32u(DisasContext *s, DisasOps *o)
2205 tcg_gen_qemu_ld32u(o->out, o->in2, get_mem_index(s));
2209 static ExitStatus op_ld64(DisasContext *s, DisasOps *o)
2211 tcg_gen_qemu_ld64(o->out, o->in2, get_mem_index(s));
2215 #ifndef CONFIG_USER_ONLY
2216 static ExitStatus op_lctl(DisasContext *s, DisasOps *o)
2218 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2219 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2220 check_privileged(s);
2221 potential_page_fault(s);
2222 gen_helper_lctl(cpu_env, r1, o->in2, r3);
2223 tcg_temp_free_i32(r1);
2224 tcg_temp_free_i32(r3);
2228 static ExitStatus op_lctlg(DisasContext *s, DisasOps *o)
2230 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2231 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2232 check_privileged(s);
2233 potential_page_fault(s);
2234 gen_helper_lctlg(cpu_env, r1, o->in2, r3);
2235 tcg_temp_free_i32(r1);
2236 tcg_temp_free_i32(r3);
2239 static ExitStatus op_lra(DisasContext *s, DisasOps *o)
2241 check_privileged(s);
2242 potential_page_fault(s);
2243 gen_helper_lra(o->out, cpu_env, o->in2);
2248 static ExitStatus op_lpsw(DisasContext *s, DisasOps *o)
2252 check_privileged(s);
2254 t1 = tcg_temp_new_i64();
2255 t2 = tcg_temp_new_i64();
2256 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2257 tcg_gen_addi_i64(o->in2, o->in2, 4);
2258 tcg_gen_qemu_ld32u(t2, o->in2, get_mem_index(s));
2259 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2260 tcg_gen_shli_i64(t1, t1, 32);
2261 gen_helper_load_psw(cpu_env, t1, t2);
2262 tcg_temp_free_i64(t1);
2263 tcg_temp_free_i64(t2);
2264 return EXIT_NORETURN;
2268 static ExitStatus op_lam(DisasContext *s, DisasOps *o)
2270 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2271 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2272 potential_page_fault(s);
2273 gen_helper_lam(cpu_env, r1, o->in2, r3);
2274 tcg_temp_free_i32(r1);
2275 tcg_temp_free_i32(r3);
2279 static ExitStatus op_lm32(DisasContext *s, DisasOps *o)
2281 int r1 = get_field(s->fields, r1);
2282 int r3 = get_field(s->fields, r3);
2283 TCGv_i64 t = tcg_temp_new_i64();
2284 TCGv_i64 t4 = tcg_const_i64(4);
2287 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2288 store_reg32_i64(r1, t);
2292 tcg_gen_add_i64(o->in2, o->in2, t4);
2296 tcg_temp_free_i64(t);
2297 tcg_temp_free_i64(t4);
2301 static ExitStatus op_lmh(DisasContext *s, DisasOps *o)
2303 int r1 = get_field(s->fields, r1);
2304 int r3 = get_field(s->fields, r3);
2305 TCGv_i64 t = tcg_temp_new_i64();
2306 TCGv_i64 t4 = tcg_const_i64(4);
2309 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2310 store_reg32h_i64(r1, t);
2314 tcg_gen_add_i64(o->in2, o->in2, t4);
2318 tcg_temp_free_i64(t);
2319 tcg_temp_free_i64(t4);
2323 static ExitStatus op_lm64(DisasContext *s, DisasOps *o)
2325 int r1 = get_field(s->fields, r1);
2326 int r3 = get_field(s->fields, r3);
2327 TCGv_i64 t8 = tcg_const_i64(8);
2330 tcg_gen_qemu_ld64(regs[r1], o->in2, get_mem_index(s));
2334 tcg_gen_add_i64(o->in2, o->in2, t8);
2338 tcg_temp_free_i64(t8);
2342 static ExitStatus op_mov2(DisasContext *s, DisasOps *o)
2345 o->g_out = o->g_in2;
2346 TCGV_UNUSED_I64(o->in2);
2351 static ExitStatus op_movx(DisasContext *s, DisasOps *o)
2355 o->g_out = o->g_in1;
2356 o->g_out2 = o->g_in2;
2357 TCGV_UNUSED_I64(o->in1);
2358 TCGV_UNUSED_I64(o->in2);
2359 o->g_in1 = o->g_in2 = false;
2363 static ExitStatus op_mvc(DisasContext *s, DisasOps *o)
2365 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2366 potential_page_fault(s);
2367 gen_helper_mvc(cpu_env, l, o->addr1, o->in2);
2368 tcg_temp_free_i32(l);
2372 static ExitStatus op_mvcl(DisasContext *s, DisasOps *o)
2374 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2375 TCGv_i32 r2 = tcg_const_i32(get_field(s->fields, r2));
2376 potential_page_fault(s);
2377 gen_helper_mvcl(cc_op, cpu_env, r1, r2);
2378 tcg_temp_free_i32(r1);
2379 tcg_temp_free_i32(r2);
2384 static ExitStatus op_mvcle(DisasContext *s, DisasOps *o)
2386 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2387 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2388 potential_page_fault(s);
2389 gen_helper_mvcle(cc_op, cpu_env, r1, o->in2, r3);
2390 tcg_temp_free_i32(r1);
2391 tcg_temp_free_i32(r3);
2396 #ifndef CONFIG_USER_ONLY
2397 static ExitStatus op_mvcp(DisasContext *s, DisasOps *o)
2399 int r1 = get_field(s->fields, l1);
2400 check_privileged(s);
2401 potential_page_fault(s);
2402 gen_helper_mvcp(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2407 static ExitStatus op_mvcs(DisasContext *s, DisasOps *o)
2409 int r1 = get_field(s->fields, l1);
2410 check_privileged(s);
2411 potential_page_fault(s);
2412 gen_helper_mvcs(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2418 static ExitStatus op_mvpg(DisasContext *s, DisasOps *o)
2420 potential_page_fault(s);
2421 gen_helper_mvpg(cpu_env, regs[0], o->in1, o->in2);
2426 static ExitStatus op_mvst(DisasContext *s, DisasOps *o)
2428 potential_page_fault(s);
2429 gen_helper_mvst(o->in1, cpu_env, regs[0], o->in1, o->in2);
2431 return_low128(o->in2);
2435 static ExitStatus op_mul(DisasContext *s, DisasOps *o)
2437 tcg_gen_mul_i64(o->out, o->in1, o->in2);
2441 static ExitStatus op_mul128(DisasContext *s, DisasOps *o)
2443 gen_helper_mul128(o->out, cpu_env, o->in1, o->in2);
2444 return_low128(o->out2);
2448 static ExitStatus op_meeb(DisasContext *s, DisasOps *o)
2450 gen_helper_meeb(o->out, cpu_env, o->in1, o->in2);
2454 static ExitStatus op_mdeb(DisasContext *s, DisasOps *o)
2456 gen_helper_mdeb(o->out, cpu_env, o->in1, o->in2);
2460 static ExitStatus op_mdb(DisasContext *s, DisasOps *o)
2462 gen_helper_mdb(o->out, cpu_env, o->in1, o->in2);
2466 static ExitStatus op_mxb(DisasContext *s, DisasOps *o)
2468 gen_helper_mxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2469 return_low128(o->out2);
2473 static ExitStatus op_mxdb(DisasContext *s, DisasOps *o)
2475 gen_helper_mxdb(o->out, cpu_env, o->out, o->out2, o->in2);
2476 return_low128(o->out2);
2480 static ExitStatus op_maeb(DisasContext *s, DisasOps *o)
2482 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2483 gen_helper_maeb(o->out, cpu_env, o->in1, o->in2, r3);
2484 tcg_temp_free_i64(r3);
2488 static ExitStatus op_madb(DisasContext *s, DisasOps *o)
2490 int r3 = get_field(s->fields, r3);
2491 gen_helper_madb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2495 static ExitStatus op_mseb(DisasContext *s, DisasOps *o)
2497 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2498 gen_helper_mseb(o->out, cpu_env, o->in1, o->in2, r3);
2499 tcg_temp_free_i64(r3);
2503 static ExitStatus op_msdb(DisasContext *s, DisasOps *o)
2505 int r3 = get_field(s->fields, r3);
2506 gen_helper_msdb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2510 static ExitStatus op_nabs(DisasContext *s, DisasOps *o)
2512 gen_helper_nabs_i64(o->out, o->in2);
2516 static ExitStatus op_nabsf32(DisasContext *s, DisasOps *o)
2518 tcg_gen_ori_i64(o->out, o->in2, 0x80000000ull);
2522 static ExitStatus op_nabsf64(DisasContext *s, DisasOps *o)
2524 tcg_gen_ori_i64(o->out, o->in2, 0x8000000000000000ull);
2528 static ExitStatus op_nabsf128(DisasContext *s, DisasOps *o)
2530 tcg_gen_ori_i64(o->out, o->in1, 0x8000000000000000ull);
2531 tcg_gen_mov_i64(o->out2, o->in2);
2535 static ExitStatus op_nc(DisasContext *s, DisasOps *o)
2537 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2538 potential_page_fault(s);
2539 gen_helper_nc(cc_op, cpu_env, l, o->addr1, o->in2);
2540 tcg_temp_free_i32(l);
2545 static ExitStatus op_neg(DisasContext *s, DisasOps *o)
2547 tcg_gen_neg_i64(o->out, o->in2);
2551 static ExitStatus op_negf32(DisasContext *s, DisasOps *o)
2553 tcg_gen_xori_i64(o->out, o->in2, 0x80000000ull);
2557 static ExitStatus op_negf64(DisasContext *s, DisasOps *o)
2559 tcg_gen_xori_i64(o->out, o->in2, 0x8000000000000000ull);
2563 static ExitStatus op_negf128(DisasContext *s, DisasOps *o)
2565 tcg_gen_xori_i64(o->out, o->in1, 0x8000000000000000ull);
2566 tcg_gen_mov_i64(o->out2, o->in2);
2570 static ExitStatus op_oc(DisasContext *s, DisasOps *o)
2572 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2573 potential_page_fault(s);
2574 gen_helper_oc(cc_op, cpu_env, l, o->addr1, o->in2);
2575 tcg_temp_free_i32(l);
2580 static ExitStatus op_or(DisasContext *s, DisasOps *o)
2582 tcg_gen_or_i64(o->out, o->in1, o->in2);
2586 static ExitStatus op_ori(DisasContext *s, DisasOps *o)
2588 int shift = s->insn->data & 0xff;
2589 int size = s->insn->data >> 8;
2590 uint64_t mask = ((1ull << size) - 1) << shift;
2593 tcg_gen_shli_i64(o->in2, o->in2, shift);
2594 tcg_gen_or_i64(o->out, o->in1, o->in2);
2596 /* Produce the CC from only the bits manipulated. */
2597 tcg_gen_andi_i64(cc_dst, o->out, mask);
2598 set_cc_nz_u64(s, cc_dst);
2602 #ifndef CONFIG_USER_ONLY
2603 static ExitStatus op_ptlb(DisasContext *s, DisasOps *o)
2605 check_privileged(s);
2606 gen_helper_ptlb(cpu_env);
2611 static ExitStatus op_rev16(DisasContext *s, DisasOps *o)
2613 tcg_gen_bswap16_i64(o->out, o->in2);
2617 static ExitStatus op_rev32(DisasContext *s, DisasOps *o)
2619 tcg_gen_bswap32_i64(o->out, o->in2);
2623 static ExitStatus op_rev64(DisasContext *s, DisasOps *o)
2625 tcg_gen_bswap64_i64(o->out, o->in2);
2629 static ExitStatus op_rll32(DisasContext *s, DisasOps *o)
2631 TCGv_i32 t1 = tcg_temp_new_i32();
2632 TCGv_i32 t2 = tcg_temp_new_i32();
2633 TCGv_i32 to = tcg_temp_new_i32();
2634 tcg_gen_trunc_i64_i32(t1, o->in1);
2635 tcg_gen_trunc_i64_i32(t2, o->in2);
2636 tcg_gen_rotl_i32(to, t1, t2);
2637 tcg_gen_extu_i32_i64(o->out, to);
2638 tcg_temp_free_i32(t1);
2639 tcg_temp_free_i32(t2);
2640 tcg_temp_free_i32(to);
2644 static ExitStatus op_rll64(DisasContext *s, DisasOps *o)
2646 tcg_gen_rotl_i64(o->out, o->in1, o->in2);
2650 #ifndef CONFIG_USER_ONLY
2651 static ExitStatus op_rrbe(DisasContext *s, DisasOps *o)
2653 check_privileged(s);
2654 gen_helper_rrbe(cc_op, cpu_env, o->in2);
2659 static ExitStatus op_sacf(DisasContext *s, DisasOps *o)
2661 check_privileged(s);
2662 gen_helper_sacf(cpu_env, o->in2);
2663 /* Addressing mode has changed, so end the block. */
2664 return EXIT_PC_STALE;
2668 static ExitStatus op_sar(DisasContext *s, DisasOps *o)
2670 int r1 = get_field(s->fields, r1);
2671 tcg_gen_st32_i64(o->in2, cpu_env, offsetof(CPUS390XState, aregs[r1]));
2675 static ExitStatus op_seb(DisasContext *s, DisasOps *o)
2677 gen_helper_seb(o->out, cpu_env, o->in1, o->in2);
2681 static ExitStatus op_sdb(DisasContext *s, DisasOps *o)
2683 gen_helper_sdb(o->out, cpu_env, o->in1, o->in2);
2687 static ExitStatus op_sxb(DisasContext *s, DisasOps *o)
2689 gen_helper_sxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2690 return_low128(o->out2);
2694 static ExitStatus op_sqeb(DisasContext *s, DisasOps *o)
2696 gen_helper_sqeb(o->out, cpu_env, o->in2);
2700 static ExitStatus op_sqdb(DisasContext *s, DisasOps *o)
2702 gen_helper_sqdb(o->out, cpu_env, o->in2);
2706 static ExitStatus op_sqxb(DisasContext *s, DisasOps *o)
2708 gen_helper_sqxb(o->out, cpu_env, o->in1, o->in2);
2709 return_low128(o->out2);
2713 #ifndef CONFIG_USER_ONLY
2714 static ExitStatus op_sigp(DisasContext *s, DisasOps *o)
2716 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2717 check_privileged(s);
2718 potential_page_fault(s);
2719 gen_helper_sigp(cc_op, cpu_env, o->in2, r1, o->in1);
2720 tcg_temp_free_i32(r1);
2725 static ExitStatus op_sla(DisasContext *s, DisasOps *o)
2727 uint64_t sign = 1ull << s->insn->data;
2728 enum cc_op cco = s->insn->data == 31 ? CC_OP_SLA_32 : CC_OP_SLA_64;
2729 gen_op_update2_cc_i64(s, cco, o->in1, o->in2);
2730 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2731 /* The arithmetic left shift is curious in that it does not affect
2732 the sign bit. Copy that over from the source unchanged. */
2733 tcg_gen_andi_i64(o->out, o->out, ~sign);
2734 tcg_gen_andi_i64(o->in1, o->in1, sign);
2735 tcg_gen_or_i64(o->out, o->out, o->in1);
2739 static ExitStatus op_sll(DisasContext *s, DisasOps *o)
2741 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2745 static ExitStatus op_sra(DisasContext *s, DisasOps *o)
2747 tcg_gen_sar_i64(o->out, o->in1, o->in2);
2751 static ExitStatus op_srl(DisasContext *s, DisasOps *o)
2753 tcg_gen_shr_i64(o->out, o->in1, o->in2);
2757 static ExitStatus op_sfpc(DisasContext *s, DisasOps *o)
2759 gen_helper_sfpc(cpu_env, o->in2);
2763 #ifndef CONFIG_USER_ONLY
2764 static ExitStatus op_spka(DisasContext *s, DisasOps *o)
2766 check_privileged(s);
2767 tcg_gen_shri_i64(o->in2, o->in2, 4);
2768 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, PSW_SHIFT_KEY - 4, 4);
2772 static ExitStatus op_sske(DisasContext *s, DisasOps *o)
2774 check_privileged(s);
2775 gen_helper_sske(cpu_env, o->in1, o->in2);
2779 static ExitStatus op_ssm(DisasContext *s, DisasOps *o)
2781 check_privileged(s);
2782 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, 56, 8);
2786 static ExitStatus op_stap(DisasContext *s, DisasOps *o)
2788 check_privileged(s);
2789 /* ??? Surely cpu address != cpu number. In any case the previous
2790 version of this stored more than the required half-word, so it
2791 is unlikely this has ever been tested. */
2792 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
2796 static ExitStatus op_stck(DisasContext *s, DisasOps *o)
2798 gen_helper_stck(o->out, cpu_env);
2799 /* ??? We don't implement clock states. */
2800 gen_op_movi_cc(s, 0);
2804 static ExitStatus op_stcke(DisasContext *s, DisasOps *o)
2806 TCGv_i64 c1 = tcg_temp_new_i64();
2807 TCGv_i64 c2 = tcg_temp_new_i64();
2808 gen_helper_stck(c1, cpu_env);
2809 /* Shift the 64-bit value into its place as a zero-extended
2810 104-bit value. Note that "bit positions 64-103 are always
2811 non-zero so that they compare differently to STCK"; we set
2812 the least significant bit to 1. */
2813 tcg_gen_shli_i64(c2, c1, 56);
2814 tcg_gen_shri_i64(c1, c1, 8);
2815 tcg_gen_ori_i64(c2, c2, 0x10000);
2816 tcg_gen_qemu_st64(c1, o->in2, get_mem_index(s));
2817 tcg_gen_addi_i64(o->in2, o->in2, 8);
2818 tcg_gen_qemu_st64(c2, o->in2, get_mem_index(s));
2819 tcg_temp_free_i64(c1);
2820 tcg_temp_free_i64(c2);
2821 /* ??? We don't implement clock states. */
2822 gen_op_movi_cc(s, 0);
2826 static ExitStatus op_sckc(DisasContext *s, DisasOps *o)
2828 check_privileged(s);
2829 gen_helper_sckc(cpu_env, o->in2);
2833 static ExitStatus op_stckc(DisasContext *s, DisasOps *o)
2835 check_privileged(s);
2836 gen_helper_stckc(o->out, cpu_env);
2840 static ExitStatus op_stctg(DisasContext *s, DisasOps *o)
2842 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2843 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2844 check_privileged(s);
2845 potential_page_fault(s);
2846 gen_helper_stctg(cpu_env, r1, o->in2, r3);
2847 tcg_temp_free_i32(r1);
2848 tcg_temp_free_i32(r3);
2852 static ExitStatus op_stctl(DisasContext *s, DisasOps *o)
2854 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2855 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2856 check_privileged(s);
2857 potential_page_fault(s);
2858 gen_helper_stctl(cpu_env, r1, o->in2, r3);
2859 tcg_temp_free_i32(r1);
2860 tcg_temp_free_i32(r3);
2864 static ExitStatus op_stidp(DisasContext *s, DisasOps *o)
2866 check_privileged(s);
2867 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
2871 static ExitStatus op_spt(DisasContext *s, DisasOps *o)
2873 check_privileged(s);
2874 gen_helper_spt(cpu_env, o->in2);
2878 static ExitStatus op_stpt(DisasContext *s, DisasOps *o)
2880 check_privileged(s);
2881 gen_helper_stpt(o->out, cpu_env);
2885 static ExitStatus op_stsi(DisasContext *s, DisasOps *o)
2887 check_privileged(s);
2888 potential_page_fault(s);
2889 gen_helper_stsi(cc_op, cpu_env, o->in2, regs[0], regs[1]);
2894 static ExitStatus op_spx(DisasContext *s, DisasOps *o)
2896 check_privileged(s);
2897 gen_helper_spx(cpu_env, o->in2);
2901 static ExitStatus op_subchannel(DisasContext *s, DisasOps *o)
2903 check_privileged(s);
2904 /* Not operational. */
2905 gen_op_movi_cc(s, 3);
2909 static ExitStatus op_stpx(DisasContext *s, DisasOps *o)
2911 check_privileged(s);
2912 tcg_gen_ld_i64(o->out, cpu_env, offsetof(CPUS390XState, psa));
2913 tcg_gen_andi_i64(o->out, o->out, 0x7fffe000);
2917 static ExitStatus op_stnosm(DisasContext *s, DisasOps *o)
2919 uint64_t i2 = get_field(s->fields, i2);
2922 check_privileged(s);
2924 /* It is important to do what the instruction name says: STORE THEN.
2925 If we let the output hook perform the store then if we fault and
2926 restart, we'll have the wrong SYSTEM MASK in place. */
2927 t = tcg_temp_new_i64();
2928 tcg_gen_shri_i64(t, psw_mask, 56);
2929 tcg_gen_qemu_st8(t, o->addr1, get_mem_index(s));
2930 tcg_temp_free_i64(t);
2932 if (s->fields->op == 0xac) {
2933 tcg_gen_andi_i64(psw_mask, psw_mask,
2934 (i2 << 56) | 0x00ffffffffffffffull);
2936 tcg_gen_ori_i64(psw_mask, psw_mask, i2 << 56);
2941 static ExitStatus op_stura(DisasContext *s, DisasOps *o)
2943 check_privileged(s);
2944 potential_page_fault(s);
2945 gen_helper_stura(cpu_env, o->in2, o->in1);
2950 static ExitStatus op_st8(DisasContext *s, DisasOps *o)
2952 tcg_gen_qemu_st8(o->in1, o->in2, get_mem_index(s));
2956 static ExitStatus op_st16(DisasContext *s, DisasOps *o)
2958 tcg_gen_qemu_st16(o->in1, o->in2, get_mem_index(s));
2962 static ExitStatus op_st32(DisasContext *s, DisasOps *o)
2964 tcg_gen_qemu_st32(o->in1, o->in2, get_mem_index(s));
2968 static ExitStatus op_st64(DisasContext *s, DisasOps *o)
2970 tcg_gen_qemu_st64(o->in1, o->in2, get_mem_index(s));
2974 static ExitStatus op_stam(DisasContext *s, DisasOps *o)
2976 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2977 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2978 potential_page_fault(s);
2979 gen_helper_stam(cpu_env, r1, o->in2, r3);
2980 tcg_temp_free_i32(r1);
2981 tcg_temp_free_i32(r3);
2985 static ExitStatus op_stcm(DisasContext *s, DisasOps *o)
2987 int m3 = get_field(s->fields, m3);
2988 int pos, base = s->insn->data;
2989 TCGv_i64 tmp = tcg_temp_new_i64();
2991 pos = base + ctz32(m3) * 8;
2994 /* Effectively a 32-bit store. */
2995 tcg_gen_shri_i64(tmp, o->in1, pos);
2996 tcg_gen_qemu_st32(tmp, o->in2, get_mem_index(s));
3002 /* Effectively a 16-bit store. */
3003 tcg_gen_shri_i64(tmp, o->in1, pos);
3004 tcg_gen_qemu_st16(tmp, o->in2, get_mem_index(s));
3011 /* Effectively an 8-bit store. */
3012 tcg_gen_shri_i64(tmp, o->in1, pos);
3013 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3017 /* This is going to be a sequence of shifts and stores. */
3018 pos = base + 32 - 8;
3021 tcg_gen_shri_i64(tmp, o->in1, pos);
3022 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3023 tcg_gen_addi_i64(o->in2, o->in2, 1);
3025 m3 = (m3 << 1) & 0xf;
3030 tcg_temp_free_i64(tmp);
3034 static ExitStatus op_stm(DisasContext *s, DisasOps *o)
3036 int r1 = get_field(s->fields, r1);
3037 int r3 = get_field(s->fields, r3);
3038 int size = s->insn->data;
3039 TCGv_i64 tsize = tcg_const_i64(size);
3043 tcg_gen_qemu_st64(regs[r1], o->in2, get_mem_index(s));
3045 tcg_gen_qemu_st32(regs[r1], o->in2, get_mem_index(s));
3050 tcg_gen_add_i64(o->in2, o->in2, tsize);
3054 tcg_temp_free_i64(tsize);
3058 static ExitStatus op_stmh(DisasContext *s, DisasOps *o)
3060 int r1 = get_field(s->fields, r1);
3061 int r3 = get_field(s->fields, r3);
3062 TCGv_i64 t = tcg_temp_new_i64();
3063 TCGv_i64 t4 = tcg_const_i64(4);
3064 TCGv_i64 t32 = tcg_const_i64(32);
3067 tcg_gen_shl_i64(t, regs[r1], t32);
3068 tcg_gen_qemu_st32(t, o->in2, get_mem_index(s));
3072 tcg_gen_add_i64(o->in2, o->in2, t4);
3076 tcg_temp_free_i64(t);
3077 tcg_temp_free_i64(t4);
3078 tcg_temp_free_i64(t32);
3082 static ExitStatus op_srst(DisasContext *s, DisasOps *o)
3084 potential_page_fault(s);
3085 gen_helper_srst(o->in1, cpu_env, regs[0], o->in1, o->in2);
3087 return_low128(o->in2);
3091 static ExitStatus op_sub(DisasContext *s, DisasOps *o)
3093 tcg_gen_sub_i64(o->out, o->in1, o->in2);
3097 static ExitStatus op_subb(DisasContext *s, DisasOps *o)
3102 tcg_gen_not_i64(o->in2, o->in2);
3103 tcg_gen_add_i64(o->out, o->in1, o->in2);
3105 /* XXX possible optimization point */
3107 cc = tcg_temp_new_i64();
3108 tcg_gen_extu_i32_i64(cc, cc_op);
3109 tcg_gen_shri_i64(cc, cc, 1);
3110 tcg_gen_add_i64(o->out, o->out, cc);
3111 tcg_temp_free_i64(cc);
3115 static ExitStatus op_svc(DisasContext *s, DisasOps *o)
3122 t = tcg_const_i32(get_field(s->fields, i1) & 0xff);
3123 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_code));
3124 tcg_temp_free_i32(t);
3126 t = tcg_const_i32(s->next_pc - s->pc);
3127 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_ilen));
3128 tcg_temp_free_i32(t);
3130 gen_exception(EXCP_SVC);
3131 return EXIT_NORETURN;
3134 static ExitStatus op_tceb(DisasContext *s, DisasOps *o)
3136 gen_helper_tceb(cc_op, o->in1, o->in2);
3141 static ExitStatus op_tcdb(DisasContext *s, DisasOps *o)
3143 gen_helper_tcdb(cc_op, o->in1, o->in2);
3148 static ExitStatus op_tcxb(DisasContext *s, DisasOps *o)
3150 gen_helper_tcxb(cc_op, o->out, o->out2, o->in2);
3155 #ifndef CONFIG_USER_ONLY
3156 static ExitStatus op_tprot(DisasContext *s, DisasOps *o)
3158 potential_page_fault(s);
3159 gen_helper_tprot(cc_op, o->addr1, o->in2);
3165 static ExitStatus op_tr(DisasContext *s, DisasOps *o)
3167 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3168 potential_page_fault(s);
3169 gen_helper_tr(cpu_env, l, o->addr1, o->in2);
3170 tcg_temp_free_i32(l);
3175 static ExitStatus op_unpk(DisasContext *s, DisasOps *o)
3177 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3178 potential_page_fault(s);
3179 gen_helper_unpk(cpu_env, l, o->addr1, o->in2);
3180 tcg_temp_free_i32(l);
3184 static ExitStatus op_xc(DisasContext *s, DisasOps *o)
3186 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3187 potential_page_fault(s);
3188 gen_helper_xc(cc_op, cpu_env, l, o->addr1, o->in2);
3189 tcg_temp_free_i32(l);
3194 static ExitStatus op_xor(DisasContext *s, DisasOps *o)
3196 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3200 static ExitStatus op_xori(DisasContext *s, DisasOps *o)
3202 int shift = s->insn->data & 0xff;
3203 int size = s->insn->data >> 8;
3204 uint64_t mask = ((1ull << size) - 1) << shift;
3207 tcg_gen_shli_i64(o->in2, o->in2, shift);
3208 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3210 /* Produce the CC from only the bits manipulated. */
3211 tcg_gen_andi_i64(cc_dst, o->out, mask);
3212 set_cc_nz_u64(s, cc_dst);
3216 static ExitStatus op_zero(DisasContext *s, DisasOps *o)
3218 o->out = tcg_const_i64(0);
3222 static ExitStatus op_zero2(DisasContext *s, DisasOps *o)
3224 o->out = tcg_const_i64(0);
3230 /* ====================================================================== */
3231 /* The "Cc OUTput" generators. Given the generated output (and in some cases
3232 the original inputs), update the various cc data structures in order to
3233 be able to compute the new condition code. */
3235 static void cout_abs32(DisasContext *s, DisasOps *o)
3237 gen_op_update1_cc_i64(s, CC_OP_ABS_32, o->out);
3240 static void cout_abs64(DisasContext *s, DisasOps *o)
3242 gen_op_update1_cc_i64(s, CC_OP_ABS_64, o->out);
3245 static void cout_adds32(DisasContext *s, DisasOps *o)
3247 gen_op_update3_cc_i64(s, CC_OP_ADD_32, o->in1, o->in2, o->out);
3250 static void cout_adds64(DisasContext *s, DisasOps *o)
3252 gen_op_update3_cc_i64(s, CC_OP_ADD_64, o->in1, o->in2, o->out);
3255 static void cout_addu32(DisasContext *s, DisasOps *o)
3257 gen_op_update3_cc_i64(s, CC_OP_ADDU_32, o->in1, o->in2, o->out);
3260 static void cout_addu64(DisasContext *s, DisasOps *o)
3262 gen_op_update3_cc_i64(s, CC_OP_ADDU_64, o->in1, o->in2, o->out);
3265 static void cout_addc32(DisasContext *s, DisasOps *o)
3267 gen_op_update3_cc_i64(s, CC_OP_ADDC_32, o->in1, o->in2, o->out);
3270 static void cout_addc64(DisasContext *s, DisasOps *o)
3272 gen_op_update3_cc_i64(s, CC_OP_ADDC_64, o->in1, o->in2, o->out);
3275 static void cout_cmps32(DisasContext *s, DisasOps *o)
3277 gen_op_update2_cc_i64(s, CC_OP_LTGT_32, o->in1, o->in2);
3280 static void cout_cmps64(DisasContext *s, DisasOps *o)
3282 gen_op_update2_cc_i64(s, CC_OP_LTGT_64, o->in1, o->in2);
3285 static void cout_cmpu32(DisasContext *s, DisasOps *o)
3287 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_32, o->in1, o->in2);
3290 static void cout_cmpu64(DisasContext *s, DisasOps *o)
3292 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, o->in1, o->in2);
3295 static void cout_f32(DisasContext *s, DisasOps *o)
3297 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, o->out);
3300 static void cout_f64(DisasContext *s, DisasOps *o)
3302 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, o->out);
3305 static void cout_f128(DisasContext *s, DisasOps *o)
3307 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, o->out, o->out2);
3310 static void cout_nabs32(DisasContext *s, DisasOps *o)
3312 gen_op_update1_cc_i64(s, CC_OP_NABS_32, o->out);
3315 static void cout_nabs64(DisasContext *s, DisasOps *o)
3317 gen_op_update1_cc_i64(s, CC_OP_NABS_64, o->out);
3320 static void cout_neg32(DisasContext *s, DisasOps *o)
3322 gen_op_update1_cc_i64(s, CC_OP_COMP_32, o->out);
3325 static void cout_neg64(DisasContext *s, DisasOps *o)
3327 gen_op_update1_cc_i64(s, CC_OP_COMP_64, o->out);
3330 static void cout_nz32(DisasContext *s, DisasOps *o)
3332 tcg_gen_ext32u_i64(cc_dst, o->out);
3333 gen_op_update1_cc_i64(s, CC_OP_NZ, cc_dst);
3336 static void cout_nz64(DisasContext *s, DisasOps *o)
3338 gen_op_update1_cc_i64(s, CC_OP_NZ, o->out);
3341 static void cout_s32(DisasContext *s, DisasOps *o)
3343 gen_op_update1_cc_i64(s, CC_OP_LTGT0_32, o->out);
3346 static void cout_s64(DisasContext *s, DisasOps *o)
3348 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, o->out);
3351 static void cout_subs32(DisasContext *s, DisasOps *o)
3353 gen_op_update3_cc_i64(s, CC_OP_SUB_32, o->in1, o->in2, o->out);
3356 static void cout_subs64(DisasContext *s, DisasOps *o)
3358 gen_op_update3_cc_i64(s, CC_OP_SUB_64, o->in1, o->in2, o->out);
3361 static void cout_subu32(DisasContext *s, DisasOps *o)
3363 gen_op_update3_cc_i64(s, CC_OP_SUBU_32, o->in1, o->in2, o->out);
3366 static void cout_subu64(DisasContext *s, DisasOps *o)
3368 gen_op_update3_cc_i64(s, CC_OP_SUBU_64, o->in1, o->in2, o->out);
3371 static void cout_subb32(DisasContext *s, DisasOps *o)
3373 gen_op_update3_cc_i64(s, CC_OP_SUBB_32, o->in1, o->in2, o->out);
3376 static void cout_subb64(DisasContext *s, DisasOps *o)
3378 gen_op_update3_cc_i64(s, CC_OP_SUBB_64, o->in1, o->in2, o->out);
3381 static void cout_tm32(DisasContext *s, DisasOps *o)
3383 gen_op_update2_cc_i64(s, CC_OP_TM_32, o->in1, o->in2);
3386 static void cout_tm64(DisasContext *s, DisasOps *o)
3388 gen_op_update2_cc_i64(s, CC_OP_TM_64, o->in1, o->in2);
3391 /* ====================================================================== */
3392 /* The "PREPeration" generators. These initialize the DisasOps.OUT fields
3393 with the TCG register to which we will write. Used in combination with
3394 the "wout" generators, in some cases we need a new temporary, and in
3395 some cases we can write to a TCG global. */
3397 static void prep_new(DisasContext *s, DisasFields *f, DisasOps *o)
3399 o->out = tcg_temp_new_i64();
3402 static void prep_new_P(DisasContext *s, DisasFields *f, DisasOps *o)
3404 o->out = tcg_temp_new_i64();
3405 o->out2 = tcg_temp_new_i64();
3408 static void prep_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3410 o->out = regs[get_field(f, r1)];
3414 static void prep_r1_P(DisasContext *s, DisasFields *f, DisasOps *o)
3416 /* ??? Specification exception: r1 must be even. */
3417 int r1 = get_field(f, r1);
3419 o->out2 = regs[(r1 + 1) & 15];
3420 o->g_out = o->g_out2 = true;
3423 static void prep_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3425 o->out = fregs[get_field(f, r1)];
3429 static void prep_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3431 /* ??? Specification exception: r1 must be < 14. */
3432 int r1 = get_field(f, r1);
3434 o->out2 = fregs[(r1 + 2) & 15];
3435 o->g_out = o->g_out2 = true;
3438 /* ====================================================================== */
3439 /* The "Write OUTput" generators. These generally perform some non-trivial
3440 copy of data to TCG globals, or to main memory. The trivial cases are
3441 generally handled by having a "prep" generator install the TCG global
3442 as the destination of the operation. */
3444 static void wout_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3446 store_reg(get_field(f, r1), o->out);
3449 static void wout_r1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3451 int r1 = get_field(f, r1);
3452 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 8);
3455 static void wout_r1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3457 int r1 = get_field(f, r1);
3458 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 16);
3461 static void wout_r1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3463 store_reg32_i64(get_field(f, r1), o->out);
3466 static void wout_r1_P32(DisasContext *s, DisasFields *f, DisasOps *o)
3468 /* ??? Specification exception: r1 must be even. */
3469 int r1 = get_field(f, r1);
3470 store_reg32_i64(r1, o->out);
3471 store_reg32_i64((r1 + 1) & 15, o->out2);
3474 static void wout_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3476 /* ??? Specification exception: r1 must be even. */
3477 int r1 = get_field(f, r1);
3478 store_reg32_i64((r1 + 1) & 15, o->out);
3479 tcg_gen_shri_i64(o->out, o->out, 32);
3480 store_reg32_i64(r1, o->out);
3483 static void wout_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3485 store_freg32_i64(get_field(f, r1), o->out);
3488 static void wout_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3490 store_freg(get_field(f, r1), o->out);
3493 static void wout_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3495 /* ??? Specification exception: r1 must be < 14. */
3496 int f1 = get_field(s->fields, r1);
3497 store_freg(f1, o->out);
3498 store_freg((f1 + 2) & 15, o->out2);
3501 static void wout_cond_r1r2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3503 if (get_field(f, r1) != get_field(f, r2)) {
3504 store_reg32_i64(get_field(f, r1), o->out);
3508 static void wout_cond_e1e2(DisasContext *s, DisasFields *f, DisasOps *o)
3510 if (get_field(f, r1) != get_field(f, r2)) {
3511 store_freg32_i64(get_field(f, r1), o->out);
3515 static void wout_m1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3517 tcg_gen_qemu_st8(o->out, o->addr1, get_mem_index(s));
3520 static void wout_m1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3522 tcg_gen_qemu_st16(o->out, o->addr1, get_mem_index(s));
3525 static void wout_m1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3527 tcg_gen_qemu_st32(o->out, o->addr1, get_mem_index(s));
3530 static void wout_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3532 tcg_gen_qemu_st64(o->out, o->addr1, get_mem_index(s));
3535 static void wout_m2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3537 tcg_gen_qemu_st32(o->out, o->in2, get_mem_index(s));
3540 /* ====================================================================== */
3541 /* The "INput 1" generators. These load the first operand to an insn. */
3543 static void in1_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3545 o->in1 = load_reg(get_field(f, r1));
3548 static void in1_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3550 o->in1 = regs[get_field(f, r1)];
3554 static void in1_r1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3556 o->in1 = tcg_temp_new_i64();
3557 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r1)]);
3560 static void in1_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3562 o->in1 = tcg_temp_new_i64();
3563 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r1)]);
3566 static void in1_r1_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
3568 o->in1 = tcg_temp_new_i64();
3569 tcg_gen_shri_i64(o->in1, regs[get_field(f, r1)], 32);
3572 static void in1_r1p1(DisasContext *s, DisasFields *f, DisasOps *o)
3574 /* ??? Specification exception: r1 must be even. */
3575 int r1 = get_field(f, r1);
3576 o->in1 = load_reg((r1 + 1) & 15);
3579 static void in1_r1p1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3581 /* ??? Specification exception: r1 must be even. */
3582 int r1 = get_field(f, r1);
3583 o->in1 = tcg_temp_new_i64();
3584 tcg_gen_ext32s_i64(o->in1, regs[(r1 + 1) & 15]);
3587 static void in1_r1p1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3589 /* ??? Specification exception: r1 must be even. */
3590 int r1 = get_field(f, r1);
3591 o->in1 = tcg_temp_new_i64();
3592 tcg_gen_ext32u_i64(o->in1, regs[(r1 + 1) & 15]);
3595 static void in1_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3597 /* ??? Specification exception: r1 must be even. */
3598 int r1 = get_field(f, r1);
3599 o->in1 = tcg_temp_new_i64();
3600 tcg_gen_concat32_i64(o->in1, regs[r1 + 1], regs[r1]);
3603 static void in1_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3605 o->in1 = load_reg(get_field(f, r2));
3608 static void in1_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3610 o->in1 = load_reg(get_field(f, r3));
3613 static void in1_r3_o(DisasContext *s, DisasFields *f, DisasOps *o)
3615 o->in1 = regs[get_field(f, r3)];
3619 static void in1_r3_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3621 o->in1 = tcg_temp_new_i64();
3622 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r3)]);
3625 static void in1_r3_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3627 o->in1 = tcg_temp_new_i64();
3628 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r3)]);
3631 static void in1_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3633 o->in1 = load_freg32_i64(get_field(f, r1));
3636 static void in1_f1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3638 o->in1 = fregs[get_field(f, r1)];
3642 static void in1_x1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3644 /* ??? Specification exception: r1 must be < 14. */
3645 int r1 = get_field(f, r1);
3647 o->out2 = fregs[(r1 + 2) & 15];
3648 o->g_out = o->g_out2 = true;
3651 static void in1_la1(DisasContext *s, DisasFields *f, DisasOps *o)
3653 o->addr1 = get_address(s, 0, get_field(f, b1), get_field(f, d1));
3656 static void in1_la2(DisasContext *s, DisasFields *f, DisasOps *o)
3658 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3659 o->addr1 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3662 static void in1_m1_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3665 o->in1 = tcg_temp_new_i64();
3666 tcg_gen_qemu_ld8u(o->in1, o->addr1, get_mem_index(s));
3669 static void in1_m1_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3672 o->in1 = tcg_temp_new_i64();
3673 tcg_gen_qemu_ld16s(o->in1, o->addr1, get_mem_index(s));
3676 static void in1_m1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3679 o->in1 = tcg_temp_new_i64();
3680 tcg_gen_qemu_ld16u(o->in1, o->addr1, get_mem_index(s));
3683 static void in1_m1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3686 o->in1 = tcg_temp_new_i64();
3687 tcg_gen_qemu_ld32s(o->in1, o->addr1, get_mem_index(s));
3690 static void in1_m1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3693 o->in1 = tcg_temp_new_i64();
3694 tcg_gen_qemu_ld32u(o->in1, o->addr1, get_mem_index(s));
3697 static void in1_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3700 o->in1 = tcg_temp_new_i64();
3701 tcg_gen_qemu_ld64(o->in1, o->addr1, get_mem_index(s));
3704 /* ====================================================================== */
3705 /* The "INput 2" generators. These load the second operand to an insn. */
3707 static void in2_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3709 o->in2 = regs[get_field(f, r1)];
3713 static void in2_r1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3715 o->in2 = tcg_temp_new_i64();
3716 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r1)]);
3719 static void in2_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3721 o->in2 = tcg_temp_new_i64();
3722 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r1)]);
3725 static void in2_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3727 o->in2 = load_reg(get_field(f, r2));
3730 static void in2_r2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3732 o->in2 = regs[get_field(f, r2)];
3736 static void in2_r2_nz(DisasContext *s, DisasFields *f, DisasOps *o)
3738 int r2 = get_field(f, r2);
3740 o->in2 = load_reg(r2);
3744 static void in2_r2_8s(DisasContext *s, DisasFields *f, DisasOps *o)
3746 o->in2 = tcg_temp_new_i64();
3747 tcg_gen_ext8s_i64(o->in2, regs[get_field(f, r2)]);
3750 static void in2_r2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3752 o->in2 = tcg_temp_new_i64();
3753 tcg_gen_ext8u_i64(o->in2, regs[get_field(f, r2)]);
3756 static void in2_r2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3758 o->in2 = tcg_temp_new_i64();
3759 tcg_gen_ext16s_i64(o->in2, regs[get_field(f, r2)]);
3762 static void in2_r2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3764 o->in2 = tcg_temp_new_i64();
3765 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r2)]);
3768 static void in2_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3770 o->in2 = load_reg(get_field(f, r3));
3773 static void in2_r2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3775 o->in2 = tcg_temp_new_i64();
3776 tcg_gen_ext32s_i64(o->in2, regs[get_field(f, r2)]);
3779 static void in2_r2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3781 o->in2 = tcg_temp_new_i64();
3782 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r2)]);
3785 static void in2_e2(DisasContext *s, DisasFields *f, DisasOps *o)
3787 o->in2 = load_freg32_i64(get_field(f, r2));
3790 static void in2_f2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3792 o->in2 = fregs[get_field(f, r2)];
3796 static void in2_x2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3798 /* ??? Specification exception: r1 must be < 14. */
3799 int r2 = get_field(f, r2);
3801 o->in2 = fregs[(r2 + 2) & 15];
3802 o->g_in1 = o->g_in2 = true;
3805 static void in2_ra2(DisasContext *s, DisasFields *f, DisasOps *o)
3807 o->in2 = get_address(s, 0, get_field(f, r2), 0);
3810 static void in2_a2(DisasContext *s, DisasFields *f, DisasOps *o)
3812 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3813 o->in2 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3816 static void in2_ri2(DisasContext *s, DisasFields *f, DisasOps *o)
3818 o->in2 = tcg_const_i64(s->pc + (int64_t)get_field(f, i2) * 2);
3821 static void in2_sh32(DisasContext *s, DisasFields *f, DisasOps *o)
3823 help_l2_shift(s, f, o, 31);
3826 static void in2_sh64(DisasContext *s, DisasFields *f, DisasOps *o)
3828 help_l2_shift(s, f, o, 63);
3831 static void in2_m2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3834 tcg_gen_qemu_ld8u(o->in2, o->in2, get_mem_index(s));
3837 static void in2_m2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3840 tcg_gen_qemu_ld16s(o->in2, o->in2, get_mem_index(s));
3843 static void in2_m2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3846 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
3849 static void in2_m2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3852 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
3855 static void in2_m2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3858 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
3861 static void in2_m2_64(DisasContext *s, DisasFields *f, DisasOps *o)
3864 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
3867 static void in2_mri2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3870 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
3873 static void in2_mri2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3876 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
3879 static void in2_mri2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3882 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
3885 static void in2_mri2_64(DisasContext *s, DisasFields *f, DisasOps *o)
3888 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
3891 static void in2_i2(DisasContext *s, DisasFields *f, DisasOps *o)
3893 o->in2 = tcg_const_i64(get_field(f, i2));
3896 static void in2_i2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3898 o->in2 = tcg_const_i64((uint8_t)get_field(f, i2));
3901 static void in2_i2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3903 o->in2 = tcg_const_i64((uint16_t)get_field(f, i2));
3906 static void in2_i2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3908 o->in2 = tcg_const_i64((uint32_t)get_field(f, i2));
3911 static void in2_i2_16u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
3913 uint64_t i2 = (uint16_t)get_field(f, i2);
3914 o->in2 = tcg_const_i64(i2 << s->insn->data);
3917 static void in2_i2_32u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
3919 uint64_t i2 = (uint32_t)get_field(f, i2);
3920 o->in2 = tcg_const_i64(i2 << s->insn->data);
3923 /* ====================================================================== */
3925 /* Find opc within the table of insns. This is formulated as a switch
3926 statement so that (1) we get compile-time notice of cut-paste errors
3927 for duplicated opcodes, and (2) the compiler generates the binary
3928 search tree, rather than us having to post-process the table. */
3930 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
3931 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
3933 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
3935 enum DisasInsnEnum {
3936 #include "insn-data.def"
3940 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
3945 .help_in1 = in1_##I1, \
3946 .help_in2 = in2_##I2, \
3947 .help_prep = prep_##P, \
3948 .help_wout = wout_##W, \
3949 .help_cout = cout_##CC, \
3950 .help_op = op_##OP, \
3954 /* Allow 0 to be used for NULL in the table below. */
3962 static const DisasInsn insn_info[] = {
3963 #include "insn-data.def"
3967 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
3968 case OPC: return &insn_info[insn_ ## NM];
3970 static const DisasInsn *lookup_opc(uint16_t opc)
3973 #include "insn-data.def"
3982 /* Extract a field from the insn. The INSN should be left-aligned in
3983 the uint64_t so that we can more easily utilize the big-bit-endian
3984 definitions we extract from the Principals of Operation. */
3986 static void extract_field(DisasFields *o, const DisasField *f, uint64_t insn)
3994 /* Zero extract the field from the insn. */
3995 r = (insn << f->beg) >> (64 - f->size);
3997 /* Sign-extend, or un-swap the field as necessary. */
3999 case 0: /* unsigned */
4001 case 1: /* signed */
4002 assert(f->size <= 32);
4003 m = 1u << (f->size - 1);
4006 case 2: /* dl+dh split, signed 20 bit. */
4007 r = ((int8_t)r << 12) | (r >> 8);
4013 /* Validate that the "compressed" encoding we selected above is valid.
4014 I.e. we havn't make two different original fields overlap. */
4015 assert(((o->presentC >> f->indexC) & 1) == 0);
4016 o->presentC |= 1 << f->indexC;
4017 o->presentO |= 1 << f->indexO;
4019 o->c[f->indexC] = r;
4022 /* Lookup the insn at the current PC, extracting the operands into O and
4023 returning the info struct for the insn. Returns NULL for invalid insn. */
4025 static const DisasInsn *extract_insn(CPUS390XState *env, DisasContext *s,
4028 uint64_t insn, pc = s->pc;
4030 const DisasInsn *info;
4032 insn = ld_code2(env, pc);
4033 op = (insn >> 8) & 0xff;
4034 ilen = get_ilen(op);
4035 s->next_pc = s->pc + ilen;
4042 insn = ld_code4(env, pc) << 32;
4045 insn = (insn << 48) | (ld_code4(env, pc + 2) << 16);
4051 /* We can't actually determine the insn format until we've looked up
4052 the full insn opcode. Which we can't do without locating the
4053 secondary opcode. Assume by default that OP2 is at bit 40; for
4054 those smaller insns that don't actually have a secondary opcode
4055 this will correctly result in OP2 = 0. */
4061 case 0xb2: /* S, RRF, RRE */
4062 case 0xb3: /* RRE, RRD, RRF */
4063 case 0xb9: /* RRE, RRF */
4064 case 0xe5: /* SSE, SIL */
4065 op2 = (insn << 8) >> 56;
4069 case 0xc0: /* RIL */
4070 case 0xc2: /* RIL */
4071 case 0xc4: /* RIL */
4072 case 0xc6: /* RIL */
4073 case 0xc8: /* SSF */
4074 case 0xcc: /* RIL */
4075 op2 = (insn << 12) >> 60;
4077 case 0xd0 ... 0xdf: /* SS */
4083 case 0xee ... 0xf3: /* SS */
4084 case 0xf8 ... 0xfd: /* SS */
4088 op2 = (insn << 40) >> 56;
4092 memset(f, 0, sizeof(*f));
4096 /* Lookup the instruction. */
4097 info = lookup_opc(op << 8 | op2);
4099 /* If we found it, extract the operands. */
4101 DisasFormat fmt = info->fmt;
4104 for (i = 0; i < NUM_C_FIELD; ++i) {
4105 extract_field(f, &format_info[fmt].op[i], insn);
4111 static ExitStatus translate_one(CPUS390XState *env, DisasContext *s)
4113 const DisasInsn *insn;
4114 ExitStatus ret = NO_EXIT;
4118 insn = extract_insn(env, s, &f);
4120 /* If not found, try the old interpreter. This includes ILLOPC. */
4122 disas_s390_insn(env, s);
4123 switch (s->is_jmp) {
4131 ret = EXIT_PC_UPDATED;
4134 ret = EXIT_NORETURN;
4144 /* Set up the strutures we use to communicate with the helpers. */
4147 o.g_out = o.g_out2 = o.g_in1 = o.g_in2 = false;
4148 TCGV_UNUSED_I64(o.out);
4149 TCGV_UNUSED_I64(o.out2);
4150 TCGV_UNUSED_I64(o.in1);
4151 TCGV_UNUSED_I64(o.in2);
4152 TCGV_UNUSED_I64(o.addr1);
4154 /* Implement the instruction. */
4155 if (insn->help_in1) {
4156 insn->help_in1(s, &f, &o);
4158 if (insn->help_in2) {
4159 insn->help_in2(s, &f, &o);
4161 if (insn->help_prep) {
4162 insn->help_prep(s, &f, &o);
4164 if (insn->help_op) {
4165 ret = insn->help_op(s, &o);
4167 if (insn->help_wout) {
4168 insn->help_wout(s, &f, &o);
4170 if (insn->help_cout) {
4171 insn->help_cout(s, &o);
4174 /* Free any temporaries created by the helpers. */
4175 if (!TCGV_IS_UNUSED_I64(o.out) && !o.g_out) {
4176 tcg_temp_free_i64(o.out);
4178 if (!TCGV_IS_UNUSED_I64(o.out2) && !o.g_out2) {
4179 tcg_temp_free_i64(o.out2);
4181 if (!TCGV_IS_UNUSED_I64(o.in1) && !o.g_in1) {
4182 tcg_temp_free_i64(o.in1);
4184 if (!TCGV_IS_UNUSED_I64(o.in2) && !o.g_in2) {
4185 tcg_temp_free_i64(o.in2);
4187 if (!TCGV_IS_UNUSED_I64(o.addr1)) {
4188 tcg_temp_free_i64(o.addr1);
4191 /* Advance to the next instruction. */
4196 static inline void gen_intermediate_code_internal(CPUS390XState *env,
4197 TranslationBlock *tb,
4201 target_ulong pc_start;
4202 uint64_t next_page_start;
4203 uint16_t *gen_opc_end;
4205 int num_insns, max_insns;
4213 if (!(tb->flags & FLAG_MASK_64)) {
4214 pc_start &= 0x7fffffff;
4219 dc.cc_op = CC_OP_DYNAMIC;
4220 do_debug = dc.singlestep_enabled = env->singlestep_enabled;
4221 dc.is_jmp = DISAS_NEXT;
4223 gen_opc_end = tcg_ctx.gen_opc_buf + OPC_MAX_SIZE;
4225 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
4228 max_insns = tb->cflags & CF_COUNT_MASK;
4229 if (max_insns == 0) {
4230 max_insns = CF_COUNT_MASK;
4237 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4241 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4244 tcg_ctx.gen_opc_pc[lj] = dc.pc;
4245 gen_opc_cc_op[lj] = dc.cc_op;
4246 tcg_ctx.gen_opc_instr_start[lj] = 1;
4247 tcg_ctx.gen_opc_icount[lj] = num_insns;
4249 if (++num_insns == max_insns && (tb->cflags & CF_LAST_IO)) {
4253 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) {
4254 tcg_gen_debug_insn_start(dc.pc);
4258 if (unlikely(!QTAILQ_EMPTY(&env->breakpoints))) {
4259 QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
4260 if (bp->pc == dc.pc) {
4261 status = EXIT_PC_STALE;
4267 if (status == NO_EXIT) {
4268 status = translate_one(env, &dc);
4271 /* If we reach a page boundary, are single stepping,
4272 or exhaust instruction count, stop generation. */
4273 if (status == NO_EXIT
4274 && (dc.pc >= next_page_start
4275 || tcg_ctx.gen_opc_ptr >= gen_opc_end
4276 || num_insns >= max_insns
4278 || env->singlestep_enabled)) {
4279 status = EXIT_PC_STALE;
4281 } while (status == NO_EXIT);
4283 if (tb->cflags & CF_LAST_IO) {
4292 update_psw_addr(&dc);
4294 case EXIT_PC_UPDATED:
4295 if (singlestep && dc.cc_op != CC_OP_DYNAMIC) {
4296 gen_op_calc_cc(&dc);
4298 /* Next TB starts off with CC_OP_DYNAMIC,
4299 so make sure the cc op type is in env */
4300 gen_op_set_cc_op(&dc);
4303 gen_exception(EXCP_DEBUG);
4305 /* Generate the return instruction */
4313 gen_icount_end(tb, num_insns);
4314 *tcg_ctx.gen_opc_ptr = INDEX_op_end;
4316 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4319 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4322 tb->size = dc.pc - pc_start;
4323 tb->icount = num_insns;
4326 #if defined(S390X_DEBUG_DISAS)
4327 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
4328 qemu_log("IN: %s\n", lookup_symbol(pc_start));
4329 log_target_disas(env, pc_start, dc.pc - pc_start, 1);
4335 void gen_intermediate_code (CPUS390XState *env, struct TranslationBlock *tb)
4337 gen_intermediate_code_internal(env, tb, 0);
4340 void gen_intermediate_code_pc (CPUS390XState *env, struct TranslationBlock *tb)
4342 gen_intermediate_code_internal(env, tb, 1);
4345 void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb, int pc_pos)
4348 env->psw.addr = tcg_ctx.gen_opc_pc[pc_pos];
4349 cc_op = gen_opc_cc_op[pc_pos];
4350 if ((cc_op != CC_OP_DYNAMIC) && (cc_op != CC_OP_STATIC)) {