4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
28 # define LOG_DISAS(...) do { } while (0)
32 #include "disas/disas.h"
35 #include "qemu/host-utils.h"
37 /* global register indexes */
38 static TCGv_ptr cpu_env;
40 #include "exec/gen-icount.h"
46 /* Information that (most) every instruction needs to manipulate. */
47 typedef struct DisasContext DisasContext;
48 typedef struct DisasInsn DisasInsn;
49 typedef struct DisasFields DisasFields;
52 struct TranslationBlock *tb;
53 const DisasInsn *insn;
57 bool singlestep_enabled;
60 /* Information carried about a condition to be evaluated. */
67 struct { TCGv_i64 a, b; } s64;
68 struct { TCGv_i32 a, b; } s32;
74 #ifdef DEBUG_INLINE_BRANCHES
75 static uint64_t inline_branch_hit[CC_OP_MAX];
76 static uint64_t inline_branch_miss[CC_OP_MAX];
79 static uint64_t pc_to_link_info(DisasContext *s, uint64_t pc)
81 if (!(s->tb->flags & FLAG_MASK_64)) {
82 if (s->tb->flags & FLAG_MASK_32) {
83 return pc | 0x80000000;
89 void cpu_dump_state(CPUS390XState *env, FILE *f, fprintf_function cpu_fprintf,
95 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %15s\n",
96 env->psw.mask, env->psw.addr, cc_name(env->cc_op));
98 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %02x\n",
99 env->psw.mask, env->psw.addr, env->cc_op);
102 for (i = 0; i < 16; i++) {
103 cpu_fprintf(f, "R%02d=%016" PRIx64, i, env->regs[i]);
105 cpu_fprintf(f, "\n");
111 for (i = 0; i < 16; i++) {
112 cpu_fprintf(f, "F%02d=%016" PRIx64, i, env->fregs[i].ll);
114 cpu_fprintf(f, "\n");
120 #ifndef CONFIG_USER_ONLY
121 for (i = 0; i < 16; i++) {
122 cpu_fprintf(f, "C%02d=%016" PRIx64, i, env->cregs[i]);
124 cpu_fprintf(f, "\n");
131 #ifdef DEBUG_INLINE_BRANCHES
132 for (i = 0; i < CC_OP_MAX; i++) {
133 cpu_fprintf(f, " %15s = %10ld\t%10ld\n", cc_name(i),
134 inline_branch_miss[i], inline_branch_hit[i]);
138 cpu_fprintf(f, "\n");
141 static TCGv_i64 psw_addr;
142 static TCGv_i64 psw_mask;
144 static TCGv_i32 cc_op;
145 static TCGv_i64 cc_src;
146 static TCGv_i64 cc_dst;
147 static TCGv_i64 cc_vr;
149 static char cpu_reg_names[32][4];
150 static TCGv_i64 regs[16];
151 static TCGv_i64 fregs[16];
153 static uint8_t gen_opc_cc_op[OPC_BUF_SIZE];
155 void s390x_translate_init(void)
159 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
160 psw_addr = tcg_global_mem_new_i64(TCG_AREG0,
161 offsetof(CPUS390XState, psw.addr),
163 psw_mask = tcg_global_mem_new_i64(TCG_AREG0,
164 offsetof(CPUS390XState, psw.mask),
167 cc_op = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUS390XState, cc_op),
169 cc_src = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_src),
171 cc_dst = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_dst),
173 cc_vr = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUS390XState, cc_vr),
176 for (i = 0; i < 16; i++) {
177 snprintf(cpu_reg_names[i], sizeof(cpu_reg_names[0]), "r%d", i);
178 regs[i] = tcg_global_mem_new(TCG_AREG0,
179 offsetof(CPUS390XState, regs[i]),
183 for (i = 0; i < 16; i++) {
184 snprintf(cpu_reg_names[i + 16], sizeof(cpu_reg_names[0]), "f%d", i);
185 fregs[i] = tcg_global_mem_new(TCG_AREG0,
186 offsetof(CPUS390XState, fregs[i].d),
187 cpu_reg_names[i + 16]);
190 /* register helpers */
195 static TCGv_i64 load_reg(int reg)
197 TCGv_i64 r = tcg_temp_new_i64();
198 tcg_gen_mov_i64(r, regs[reg]);
202 static TCGv_i64 load_freg32_i64(int reg)
204 TCGv_i64 r = tcg_temp_new_i64();
205 tcg_gen_shri_i64(r, fregs[reg], 32);
209 static void store_reg(int reg, TCGv_i64 v)
211 tcg_gen_mov_i64(regs[reg], v);
214 static void store_freg(int reg, TCGv_i64 v)
216 tcg_gen_mov_i64(fregs[reg], v);
219 static void store_reg32_i64(int reg, TCGv_i64 v)
221 /* 32 bit register writes keep the upper half */
222 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 0, 32);
225 static void store_reg32h_i64(int reg, TCGv_i64 v)
227 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 32, 32);
230 static void store_freg32_i64(int reg, TCGv_i64 v)
232 tcg_gen_deposit_i64(fregs[reg], fregs[reg], v, 32, 32);
235 static void return_low128(TCGv_i64 dest)
237 tcg_gen_ld_i64(dest, cpu_env, offsetof(CPUS390XState, retxl));
240 static void update_psw_addr(DisasContext *s)
243 tcg_gen_movi_i64(psw_addr, s->pc);
246 static void update_cc_op(DisasContext *s)
248 if (s->cc_op != CC_OP_DYNAMIC && s->cc_op != CC_OP_STATIC) {
249 tcg_gen_movi_i32(cc_op, s->cc_op);
253 static void potential_page_fault(DisasContext *s)
259 static inline uint64_t ld_code2(CPUS390XState *env, uint64_t pc)
261 return (uint64_t)cpu_lduw_code(env, pc);
264 static inline uint64_t ld_code4(CPUS390XState *env, uint64_t pc)
266 return (uint64_t)(uint32_t)cpu_ldl_code(env, pc);
269 static inline uint64_t ld_code6(CPUS390XState *env, uint64_t pc)
271 return (ld_code2(env, pc) << 32) | ld_code4(env, pc + 2);
274 static int get_mem_index(DisasContext *s)
276 switch (s->tb->flags & FLAG_MASK_ASC) {
277 case PSW_ASC_PRIMARY >> 32:
279 case PSW_ASC_SECONDARY >> 32:
281 case PSW_ASC_HOME >> 32:
289 static void gen_exception(int excp)
291 TCGv_i32 tmp = tcg_const_i32(excp);
292 gen_helper_exception(cpu_env, tmp);
293 tcg_temp_free_i32(tmp);
296 static void gen_program_exception(DisasContext *s, int code)
300 /* Remember what pgm exeption this was. */
301 tmp = tcg_const_i32(code);
302 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_code));
303 tcg_temp_free_i32(tmp);
305 tmp = tcg_const_i32(s->next_pc - s->pc);
306 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_ilen));
307 tcg_temp_free_i32(tmp);
309 /* Advance past instruction. */
316 /* Trigger exception. */
317 gen_exception(EXCP_PGM);
320 static inline void gen_illegal_opcode(DisasContext *s)
322 gen_program_exception(s, PGM_SPECIFICATION);
325 static inline void check_privileged(DisasContext *s)
327 if (s->tb->flags & (PSW_MASK_PSTATE >> 32)) {
328 gen_program_exception(s, PGM_PRIVILEGED);
332 static TCGv_i64 get_address(DisasContext *s, int x2, int b2, int d2)
336 /* 31-bitify the immediate part; register contents are dealt with below */
337 if (!(s->tb->flags & FLAG_MASK_64)) {
343 tmp = tcg_const_i64(d2);
344 tcg_gen_add_i64(tmp, tmp, regs[x2]);
349 tcg_gen_add_i64(tmp, tmp, regs[b2]);
353 tmp = tcg_const_i64(d2);
354 tcg_gen_add_i64(tmp, tmp, regs[b2]);
359 tmp = tcg_const_i64(d2);
362 /* 31-bit mode mask if there are values loaded from registers */
363 if (!(s->tb->flags & FLAG_MASK_64) && (x2 || b2)) {
364 tcg_gen_andi_i64(tmp, tmp, 0x7fffffffUL);
370 static inline void gen_op_movi_cc(DisasContext *s, uint32_t val)
372 s->cc_op = CC_OP_CONST0 + val;
375 static void gen_op_update1_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 dst)
377 tcg_gen_discard_i64(cc_src);
378 tcg_gen_mov_i64(cc_dst, dst);
379 tcg_gen_discard_i64(cc_vr);
383 static void gen_op_update2_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
386 tcg_gen_mov_i64(cc_src, src);
387 tcg_gen_mov_i64(cc_dst, dst);
388 tcg_gen_discard_i64(cc_vr);
392 static void gen_op_update3_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
393 TCGv_i64 dst, TCGv_i64 vr)
395 tcg_gen_mov_i64(cc_src, src);
396 tcg_gen_mov_i64(cc_dst, dst);
397 tcg_gen_mov_i64(cc_vr, vr);
401 static void set_cc_nz_u64(DisasContext *s, TCGv_i64 val)
403 gen_op_update1_cc_i64(s, CC_OP_NZ, val);
406 static void gen_set_cc_nz_f32(DisasContext *s, TCGv_i64 val)
408 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, val);
411 static void gen_set_cc_nz_f64(DisasContext *s, TCGv_i64 val)
413 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, val);
416 static void gen_set_cc_nz_f128(DisasContext *s, TCGv_i64 vh, TCGv_i64 vl)
418 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, vh, vl);
421 /* CC value is in env->cc_op */
422 static void set_cc_static(DisasContext *s)
424 tcg_gen_discard_i64(cc_src);
425 tcg_gen_discard_i64(cc_dst);
426 tcg_gen_discard_i64(cc_vr);
427 s->cc_op = CC_OP_STATIC;
430 /* calculates cc into cc_op */
431 static void gen_op_calc_cc(DisasContext *s)
433 TCGv_i32 local_cc_op;
436 TCGV_UNUSED_I32(local_cc_op);
437 TCGV_UNUSED_I64(dummy);
440 dummy = tcg_const_i64(0);
454 local_cc_op = tcg_const_i32(s->cc_op);
470 /* s->cc_op is the cc value */
471 tcg_gen_movi_i32(cc_op, s->cc_op - CC_OP_CONST0);
474 /* env->cc_op already is the cc value */
489 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, dummy, cc_dst, dummy);
494 case CC_OP_LTUGTU_32:
495 case CC_OP_LTUGTU_64:
502 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, dummy);
517 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, cc_vr);
520 /* unknown operation - assume 3 arguments and cc_op in env */
521 gen_helper_calc_cc(cc_op, cpu_env, cc_op, cc_src, cc_dst, cc_vr);
527 if (!TCGV_IS_UNUSED_I32(local_cc_op)) {
528 tcg_temp_free_i32(local_cc_op);
530 if (!TCGV_IS_UNUSED_I64(dummy)) {
531 tcg_temp_free_i64(dummy);
534 /* We now have cc in cc_op as constant */
538 static int use_goto_tb(DisasContext *s, uint64_t dest)
540 /* NOTE: we handle the case where the TB spans two pages here */
541 return (((dest & TARGET_PAGE_MASK) == (s->tb->pc & TARGET_PAGE_MASK)
542 || (dest & TARGET_PAGE_MASK) == ((s->pc - 1) & TARGET_PAGE_MASK))
543 && !s->singlestep_enabled
544 && !(s->tb->cflags & CF_LAST_IO));
547 static void account_noninline_branch(DisasContext *s, int cc_op)
549 #ifdef DEBUG_INLINE_BRANCHES
550 inline_branch_miss[cc_op]++;
554 static void account_inline_branch(DisasContext *s, int cc_op)
556 #ifdef DEBUG_INLINE_BRANCHES
557 inline_branch_hit[cc_op]++;
561 /* Table of mask values to comparison codes, given a comparison as input.
562 For a true comparison CC=3 will never be set, but we treat this
563 conservatively for possible use when CC=3 indicates overflow. */
564 static const TCGCond ltgt_cond[16] = {
565 TCG_COND_NEVER, TCG_COND_NEVER, /* | | | x */
566 TCG_COND_GT, TCG_COND_NEVER, /* | | GT | x */
567 TCG_COND_LT, TCG_COND_NEVER, /* | LT | | x */
568 TCG_COND_NE, TCG_COND_NEVER, /* | LT | GT | x */
569 TCG_COND_EQ, TCG_COND_NEVER, /* EQ | | | x */
570 TCG_COND_GE, TCG_COND_NEVER, /* EQ | | GT | x */
571 TCG_COND_LE, TCG_COND_NEVER, /* EQ | LT | | x */
572 TCG_COND_ALWAYS, TCG_COND_ALWAYS, /* EQ | LT | GT | x */
575 /* Table of mask values to comparison codes, given a logic op as input.
576 For such, only CC=0 and CC=1 should be possible. */
577 static const TCGCond nz_cond[16] = {
579 TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER, TCG_COND_NEVER,
581 TCG_COND_NE, TCG_COND_NE, TCG_COND_NE, TCG_COND_NE,
583 TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ, TCG_COND_EQ,
584 /* EQ | NE | x | x */
585 TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS, TCG_COND_ALWAYS,
588 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
589 details required to generate a TCG comparison. */
590 static void disas_jcc(DisasContext *s, DisasCompare *c, uint32_t mask)
593 enum cc_op old_cc_op = s->cc_op;
595 if (mask == 15 || mask == 0) {
596 c->cond = (mask ? TCG_COND_ALWAYS : TCG_COND_NEVER);
599 c->g1 = c->g2 = true;
604 /* Find the TCG condition for the mask + cc op. */
610 cond = ltgt_cond[mask];
611 if (cond == TCG_COND_NEVER) {
614 account_inline_branch(s, old_cc_op);
617 case CC_OP_LTUGTU_32:
618 case CC_OP_LTUGTU_64:
619 cond = tcg_unsigned_cond(ltgt_cond[mask]);
620 if (cond == TCG_COND_NEVER) {
623 account_inline_branch(s, old_cc_op);
627 cond = nz_cond[mask];
628 if (cond == TCG_COND_NEVER) {
631 account_inline_branch(s, old_cc_op);
646 account_inline_branch(s, old_cc_op);
661 account_inline_branch(s, old_cc_op);
665 switch (mask & 0xa) {
666 case 8: /* src == 0 -> no one bit found */
669 case 2: /* src != 0 -> one bit found */
675 account_inline_branch(s, old_cc_op);
680 /* Calculate cc value. */
685 /* Jump based on CC. We'll load up the real cond below;
686 the assignment here merely avoids a compiler warning. */
687 account_noninline_branch(s, old_cc_op);
688 old_cc_op = CC_OP_STATIC;
689 cond = TCG_COND_NEVER;
693 /* Load up the arguments of the comparison. */
695 c->g1 = c->g2 = false;
699 c->u.s32.a = tcg_temp_new_i32();
700 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_dst);
701 c->u.s32.b = tcg_const_i32(0);
704 case CC_OP_LTUGTU_32:
706 c->u.s32.a = tcg_temp_new_i32();
707 tcg_gen_trunc_i64_i32(c->u.s32.a, cc_src);
708 c->u.s32.b = tcg_temp_new_i32();
709 tcg_gen_trunc_i64_i32(c->u.s32.b, cc_dst);
716 c->u.s64.b = tcg_const_i64(0);
720 case CC_OP_LTUGTU_64:
723 c->g1 = c->g2 = true;
729 c->u.s64.a = tcg_temp_new_i64();
730 c->u.s64.b = tcg_const_i64(0);
731 tcg_gen_and_i64(c->u.s64.a, cc_src, cc_dst);
739 case 0x8 | 0x4 | 0x2: /* cc != 3 */
741 c->u.s32.b = tcg_const_i32(3);
743 case 0x8 | 0x4 | 0x1: /* cc != 2 */
745 c->u.s32.b = tcg_const_i32(2);
747 case 0x8 | 0x2 | 0x1: /* cc != 1 */
749 c->u.s32.b = tcg_const_i32(1);
751 case 0x8 | 0x2: /* cc == 0 ||Â cc == 2 => (cc & 1) == 0 */
754 c->u.s32.a = tcg_temp_new_i32();
755 c->u.s32.b = tcg_const_i32(0);
756 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
758 case 0x8 | 0x4: /* cc < 2 */
760 c->u.s32.b = tcg_const_i32(2);
762 case 0x8: /* cc == 0 */
764 c->u.s32.b = tcg_const_i32(0);
766 case 0x4 | 0x2 | 0x1: /* cc != 0 */
768 c->u.s32.b = tcg_const_i32(0);
770 case 0x4 | 0x1: /* cc == 1 ||Â cc == 3 => (cc & 1) != 0 */
773 c->u.s32.a = tcg_temp_new_i32();
774 c->u.s32.b = tcg_const_i32(0);
775 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
777 case 0x4: /* cc == 1 */
779 c->u.s32.b = tcg_const_i32(1);
781 case 0x2 | 0x1: /* cc > 1 */
783 c->u.s32.b = tcg_const_i32(1);
785 case 0x2: /* cc == 2 */
787 c->u.s32.b = tcg_const_i32(2);
789 case 0x1: /* cc == 3 */
791 c->u.s32.b = tcg_const_i32(3);
794 /* CC is masked by something else: (8 >> cc) & mask. */
797 c->u.s32.a = tcg_const_i32(8);
798 c->u.s32.b = tcg_const_i32(0);
799 tcg_gen_shr_i32(c->u.s32.a, c->u.s32.a, cc_op);
800 tcg_gen_andi_i32(c->u.s32.a, c->u.s32.a, mask);
811 static void free_compare(DisasCompare *c)
815 tcg_temp_free_i64(c->u.s64.a);
817 tcg_temp_free_i32(c->u.s32.a);
822 tcg_temp_free_i64(c->u.s64.b);
824 tcg_temp_free_i32(c->u.s32.b);
829 /* ====================================================================== */
830 /* Define the insn format enumeration. */
831 #define F0(N) FMT_##N,
832 #define F1(N, X1) F0(N)
833 #define F2(N, X1, X2) F0(N)
834 #define F3(N, X1, X2, X3) F0(N)
835 #define F4(N, X1, X2, X3, X4) F0(N)
836 #define F5(N, X1, X2, X3, X4, X5) F0(N)
839 #include "insn-format.def"
849 /* Define a structure to hold the decoded fields. We'll store each inside
850 an array indexed by an enum. In order to conserve memory, we'll arrange
851 for fields that do not exist at the same time to overlap, thus the "C"
852 for compact. For checking purposes there is an "O" for original index
853 as well that will be applied to availability bitmaps. */
855 enum DisasFieldIndexO {
878 enum DisasFieldIndexC {
912 unsigned presentC:16;
913 unsigned int presentO;
917 /* This is the way fields are to be accessed out of DisasFields. */
918 #define have_field(S, F) have_field1((S), FLD_O_##F)
919 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
921 static bool have_field1(const DisasFields *f, enum DisasFieldIndexO c)
923 return (f->presentO >> c) & 1;
926 static int get_field1(const DisasFields *f, enum DisasFieldIndexO o,
927 enum DisasFieldIndexC c)
929 assert(have_field1(f, o));
933 /* Describe the layout of each field in each format. */
934 typedef struct DisasField {
938 unsigned int indexC:6;
939 enum DisasFieldIndexO indexO:8;
942 typedef struct DisasFormatInfo {
943 DisasField op[NUM_C_FIELD];
946 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
947 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
948 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
949 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
950 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
951 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
952 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
953 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
954 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
955 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
956 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
957 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
958 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
959 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
961 #define F0(N) { { } },
962 #define F1(N, X1) { { X1 } },
963 #define F2(N, X1, X2) { { X1, X2 } },
964 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
965 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
966 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
968 static const DisasFormatInfo format_info[] = {
969 #include "insn-format.def"
987 /* Generally, we'll extract operands into this structures, operate upon
988 them, and store them back. See the "in1", "in2", "prep", "wout" sets
989 of routines below for more details. */
991 bool g_out, g_out2, g_in1, g_in2;
992 TCGv_i64 out, out2, in1, in2;
996 /* Return values from translate_one, indicating the state of the TB. */
998 /* Continue the TB. */
1000 /* We have emitted one or more goto_tb. No fixup required. */
1002 /* We are not using a goto_tb (for whatever reason), but have updated
1003 the PC (for whatever reason), so there's no need to do it again on
1006 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1007 updated the PC for the next instruction to be executed. */
1009 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1010 No following code will be executed. */
1014 typedef enum DisasFacility {
1015 FAC_Z, /* zarch (default) */
1016 FAC_CASS, /* compare and swap and store */
1017 FAC_CASS2, /* compare and swap and store 2*/
1018 FAC_DFP, /* decimal floating point */
1019 FAC_DFPR, /* decimal floating point rounding */
1020 FAC_DO, /* distinct operands */
1021 FAC_EE, /* execute extensions */
1022 FAC_EI, /* extended immediate */
1023 FAC_FPE, /* floating point extension */
1024 FAC_FPSSH, /* floating point support sign handling */
1025 FAC_FPRGR, /* FPR-GR transfer */
1026 FAC_GIE, /* general instructions extension */
1027 FAC_HFP_MA, /* HFP multiply-and-add/subtract */
1028 FAC_HW, /* high-word */
1029 FAC_IEEEE_SIM, /* IEEE exception sumilation */
1030 FAC_LOC, /* load/store on condition */
1031 FAC_LD, /* long displacement */
1032 FAC_PC, /* population count */
1033 FAC_SCF, /* store clock fast */
1034 FAC_SFLE, /* store facility list extended */
1040 DisasFacility fac:6;
1044 void (*help_in1)(DisasContext *, DisasFields *, DisasOps *);
1045 void (*help_in2)(DisasContext *, DisasFields *, DisasOps *);
1046 void (*help_prep)(DisasContext *, DisasFields *, DisasOps *);
1047 void (*help_wout)(DisasContext *, DisasFields *, DisasOps *);
1048 void (*help_cout)(DisasContext *, DisasOps *);
1049 ExitStatus (*help_op)(DisasContext *, DisasOps *);
1054 /* ====================================================================== */
1055 /* Miscelaneous helpers, used by several operations. */
1057 static void help_l2_shift(DisasContext *s, DisasFields *f,
1058 DisasOps *o, int mask)
1060 int b2 = get_field(f, b2);
1061 int d2 = get_field(f, d2);
1064 o->in2 = tcg_const_i64(d2 & mask);
1066 o->in2 = get_address(s, 0, b2, d2);
1067 tcg_gen_andi_i64(o->in2, o->in2, mask);
1071 static ExitStatus help_goto_direct(DisasContext *s, uint64_t dest)
1073 if (dest == s->next_pc) {
1076 if (use_goto_tb(s, dest)) {
1079 tcg_gen_movi_i64(psw_addr, dest);
1080 tcg_gen_exit_tb((tcg_target_long)s->tb);
1081 return EXIT_GOTO_TB;
1083 tcg_gen_movi_i64(psw_addr, dest);
1084 return EXIT_PC_UPDATED;
1088 static ExitStatus help_branch(DisasContext *s, DisasCompare *c,
1089 bool is_imm, int imm, TCGv_i64 cdest)
1092 uint64_t dest = s->pc + 2 * imm;
1095 /* Take care of the special cases first. */
1096 if (c->cond == TCG_COND_NEVER) {
1101 if (dest == s->next_pc) {
1102 /* Branch to next. */
1106 if (c->cond == TCG_COND_ALWAYS) {
1107 ret = help_goto_direct(s, dest);
1111 if (TCGV_IS_UNUSED_I64(cdest)) {
1112 /* E.g. bcr %r0 -> no branch. */
1116 if (c->cond == TCG_COND_ALWAYS) {
1117 tcg_gen_mov_i64(psw_addr, cdest);
1118 ret = EXIT_PC_UPDATED;
1123 if (use_goto_tb(s, s->next_pc)) {
1124 if (is_imm && use_goto_tb(s, dest)) {
1125 /* Both exits can use goto_tb. */
1128 lab = gen_new_label();
1130 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1132 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1135 /* Branch not taken. */
1137 tcg_gen_movi_i64(psw_addr, s->next_pc);
1138 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1143 tcg_gen_movi_i64(psw_addr, dest);
1144 tcg_gen_exit_tb((tcg_target_long)s->tb + 1);
1148 /* Fallthru can use goto_tb, but taken branch cannot. */
1149 /* Store taken branch destination before the brcond. This
1150 avoids having to allocate a new local temp to hold it.
1151 We'll overwrite this in the not taken case anyway. */
1153 tcg_gen_mov_i64(psw_addr, cdest);
1156 lab = gen_new_label();
1158 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1160 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1163 /* Branch not taken. */
1166 tcg_gen_movi_i64(psw_addr, s->next_pc);
1167 tcg_gen_exit_tb((tcg_target_long)s->tb + 0);
1171 tcg_gen_movi_i64(psw_addr, dest);
1173 ret = EXIT_PC_UPDATED;
1176 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1177 Most commonly we're single-stepping or some other condition that
1178 disables all use of goto_tb. Just update the PC and exit. */
1180 TCGv_i64 next = tcg_const_i64(s->next_pc);
1182 cdest = tcg_const_i64(dest);
1186 tcg_gen_movcond_i64(c->cond, psw_addr, c->u.s64.a, c->u.s64.b,
1189 TCGv_i32 t0 = tcg_temp_new_i32();
1190 TCGv_i64 t1 = tcg_temp_new_i64();
1191 TCGv_i64 z = tcg_const_i64(0);
1192 tcg_gen_setcond_i32(c->cond, t0, c->u.s32.a, c->u.s32.b);
1193 tcg_gen_extu_i32_i64(t1, t0);
1194 tcg_temp_free_i32(t0);
1195 tcg_gen_movcond_i64(TCG_COND_NE, psw_addr, t1, z, cdest, next);
1196 tcg_temp_free_i64(t1);
1197 tcg_temp_free_i64(z);
1201 tcg_temp_free_i64(cdest);
1203 tcg_temp_free_i64(next);
1205 ret = EXIT_PC_UPDATED;
1213 /* ====================================================================== */
1214 /* The operations. These perform the bulk of the work for any insn,
1215 usually after the operands have been loaded and output initialized. */
1217 static ExitStatus op_abs(DisasContext *s, DisasOps *o)
1219 gen_helper_abs_i64(o->out, o->in2);
1223 static ExitStatus op_absf32(DisasContext *s, DisasOps *o)
1225 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffull);
1229 static ExitStatus op_absf64(DisasContext *s, DisasOps *o)
1231 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffffffffffull);
1235 static ExitStatus op_absf128(DisasContext *s, DisasOps *o)
1237 tcg_gen_andi_i64(o->out, o->in1, 0x7fffffffffffffffull);
1238 tcg_gen_mov_i64(o->out2, o->in2);
1242 static ExitStatus op_add(DisasContext *s, DisasOps *o)
1244 tcg_gen_add_i64(o->out, o->in1, o->in2);
1248 static ExitStatus op_addc(DisasContext *s, DisasOps *o)
1252 tcg_gen_add_i64(o->out, o->in1, o->in2);
1254 /* XXX possible optimization point */
1256 cc = tcg_temp_new_i64();
1257 tcg_gen_extu_i32_i64(cc, cc_op);
1258 tcg_gen_shri_i64(cc, cc, 1);
1260 tcg_gen_add_i64(o->out, o->out, cc);
1261 tcg_temp_free_i64(cc);
1265 static ExitStatus op_aeb(DisasContext *s, DisasOps *o)
1267 gen_helper_aeb(o->out, cpu_env, o->in1, o->in2);
1271 static ExitStatus op_adb(DisasContext *s, DisasOps *o)
1273 gen_helper_adb(o->out, cpu_env, o->in1, o->in2);
1277 static ExitStatus op_axb(DisasContext *s, DisasOps *o)
1279 gen_helper_axb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1280 return_low128(o->out2);
1284 static ExitStatus op_and(DisasContext *s, DisasOps *o)
1286 tcg_gen_and_i64(o->out, o->in1, o->in2);
1290 static ExitStatus op_andi(DisasContext *s, DisasOps *o)
1292 int shift = s->insn->data & 0xff;
1293 int size = s->insn->data >> 8;
1294 uint64_t mask = ((1ull << size) - 1) << shift;
1297 tcg_gen_shli_i64(o->in2, o->in2, shift);
1298 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
1299 tcg_gen_and_i64(o->out, o->in1, o->in2);
1301 /* Produce the CC from only the bits manipulated. */
1302 tcg_gen_andi_i64(cc_dst, o->out, mask);
1303 set_cc_nz_u64(s, cc_dst);
1307 static ExitStatus op_bas(DisasContext *s, DisasOps *o)
1309 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1310 if (!TCGV_IS_UNUSED_I64(o->in2)) {
1311 tcg_gen_mov_i64(psw_addr, o->in2);
1312 return EXIT_PC_UPDATED;
1318 static ExitStatus op_basi(DisasContext *s, DisasOps *o)
1320 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1321 return help_goto_direct(s, s->pc + 2 * get_field(s->fields, i2));
1324 static ExitStatus op_bc(DisasContext *s, DisasOps *o)
1326 int m1 = get_field(s->fields, m1);
1327 bool is_imm = have_field(s->fields, i2);
1328 int imm = is_imm ? get_field(s->fields, i2) : 0;
1331 disas_jcc(s, &c, m1);
1332 return help_branch(s, &c, is_imm, imm, o->in2);
1335 static ExitStatus op_bct32(DisasContext *s, DisasOps *o)
1337 int r1 = get_field(s->fields, r1);
1338 bool is_imm = have_field(s->fields, i2);
1339 int imm = is_imm ? get_field(s->fields, i2) : 0;
1343 c.cond = TCG_COND_NE;
1348 t = tcg_temp_new_i64();
1349 tcg_gen_subi_i64(t, regs[r1], 1);
1350 store_reg32_i64(r1, t);
1351 c.u.s32.a = tcg_temp_new_i32();
1352 c.u.s32.b = tcg_const_i32(0);
1353 tcg_gen_trunc_i64_i32(c.u.s32.a, t);
1354 tcg_temp_free_i64(t);
1356 return help_branch(s, &c, is_imm, imm, o->in2);
1359 static ExitStatus op_bct64(DisasContext *s, DisasOps *o)
1361 int r1 = get_field(s->fields, r1);
1362 bool is_imm = have_field(s->fields, i2);
1363 int imm = is_imm ? get_field(s->fields, i2) : 0;
1366 c.cond = TCG_COND_NE;
1371 tcg_gen_subi_i64(regs[r1], regs[r1], 1);
1372 c.u.s64.a = regs[r1];
1373 c.u.s64.b = tcg_const_i64(0);
1375 return help_branch(s, &c, is_imm, imm, o->in2);
1378 static ExitStatus op_bx32(DisasContext *s, DisasOps *o)
1380 int r1 = get_field(s->fields, r1);
1381 int r3 = get_field(s->fields, r3);
1382 bool is_imm = have_field(s->fields, i2);
1383 int imm = is_imm ? get_field(s->fields, i2) : 0;
1387 c.cond = (s->insn->data ? TCG_COND_LE : TCG_COND_GT);
1392 t = tcg_temp_new_i64();
1393 tcg_gen_add_i64(t, regs[r1], regs[r3]);
1394 c.u.s32.a = tcg_temp_new_i32();
1395 c.u.s32.b = tcg_temp_new_i32();
1396 tcg_gen_trunc_i64_i32(c.u.s32.a, t);
1397 tcg_gen_trunc_i64_i32(c.u.s32.b, regs[r3 | 1]);
1398 store_reg32_i64(r1, t);
1399 tcg_temp_free_i64(t);
1401 return help_branch(s, &c, is_imm, imm, o->in2);
1404 static ExitStatus op_bx64(DisasContext *s, DisasOps *o)
1406 int r1 = get_field(s->fields, r1);
1407 int r3 = get_field(s->fields, r3);
1408 bool is_imm = have_field(s->fields, i2);
1409 int imm = is_imm ? get_field(s->fields, i2) : 0;
1412 c.cond = (s->insn->data ? TCG_COND_LE : TCG_COND_GT);
1415 if (r1 == (r3 | 1)) {
1416 c.u.s64.b = load_reg(r3 | 1);
1419 c.u.s64.b = regs[r3 | 1];
1423 tcg_gen_add_i64(regs[r1], regs[r1], regs[r3]);
1424 c.u.s64.a = regs[r1];
1427 return help_branch(s, &c, is_imm, imm, o->in2);
1430 static ExitStatus op_cj(DisasContext *s, DisasOps *o)
1432 int imm, m3 = get_field(s->fields, m3);
1436 /* Bit 3 of the m3 field is reserved and should be zero.
1437 Choose to ignore it wrt the ltgt_cond table above. */
1438 c.cond = ltgt_cond[m3 & 14];
1439 if (s->insn->data) {
1440 c.cond = tcg_unsigned_cond(c.cond);
1442 c.is_64 = c.g1 = c.g2 = true;
1446 is_imm = have_field(s->fields, i4);
1448 imm = get_field(s->fields, i4);
1451 o->out = get_address(s, 0, get_field(s->fields, b4),
1452 get_field(s->fields, d4));
1455 return help_branch(s, &c, is_imm, imm, o->out);
1458 static ExitStatus op_ceb(DisasContext *s, DisasOps *o)
1460 gen_helper_ceb(cc_op, cpu_env, o->in1, o->in2);
1465 static ExitStatus op_cdb(DisasContext *s, DisasOps *o)
1467 gen_helper_cdb(cc_op, cpu_env, o->in1, o->in2);
1472 static ExitStatus op_cxb(DisasContext *s, DisasOps *o)
1474 gen_helper_cxb(cc_op, cpu_env, o->out, o->out2, o->in1, o->in2);
1479 static ExitStatus op_cfeb(DisasContext *s, DisasOps *o)
1481 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1482 gen_helper_cfeb(o->out, cpu_env, o->in2, m3);
1483 tcg_temp_free_i32(m3);
1484 gen_set_cc_nz_f32(s, o->in2);
1488 static ExitStatus op_cfdb(DisasContext *s, DisasOps *o)
1490 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1491 gen_helper_cfdb(o->out, cpu_env, o->in2, m3);
1492 tcg_temp_free_i32(m3);
1493 gen_set_cc_nz_f64(s, o->in2);
1497 static ExitStatus op_cfxb(DisasContext *s, DisasOps *o)
1499 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1500 gen_helper_cfxb(o->out, cpu_env, o->in1, o->in2, m3);
1501 tcg_temp_free_i32(m3);
1502 gen_set_cc_nz_f128(s, o->in1, o->in2);
1506 static ExitStatus op_cgeb(DisasContext *s, DisasOps *o)
1508 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1509 gen_helper_cgeb(o->out, cpu_env, o->in2, m3);
1510 tcg_temp_free_i32(m3);
1511 gen_set_cc_nz_f32(s, o->in2);
1515 static ExitStatus op_cgdb(DisasContext *s, DisasOps *o)
1517 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1518 gen_helper_cgdb(o->out, cpu_env, o->in2, m3);
1519 tcg_temp_free_i32(m3);
1520 gen_set_cc_nz_f64(s, o->in2);
1524 static ExitStatus op_cgxb(DisasContext *s, DisasOps *o)
1526 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1527 gen_helper_cgxb(o->out, cpu_env, o->in1, o->in2, m3);
1528 tcg_temp_free_i32(m3);
1529 gen_set_cc_nz_f128(s, o->in1, o->in2);
1533 static ExitStatus op_clfeb(DisasContext *s, DisasOps *o)
1535 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1536 gen_helper_clfeb(o->out, cpu_env, o->in2, m3);
1537 tcg_temp_free_i32(m3);
1538 gen_set_cc_nz_f32(s, o->in2);
1542 static ExitStatus op_clfdb(DisasContext *s, DisasOps *o)
1544 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1545 gen_helper_clfdb(o->out, cpu_env, o->in2, m3);
1546 tcg_temp_free_i32(m3);
1547 gen_set_cc_nz_f64(s, o->in2);
1551 static ExitStatus op_clfxb(DisasContext *s, DisasOps *o)
1553 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1554 gen_helper_clfxb(o->out, cpu_env, o->in1, o->in2, m3);
1555 tcg_temp_free_i32(m3);
1556 gen_set_cc_nz_f128(s, o->in1, o->in2);
1560 static ExitStatus op_clgeb(DisasContext *s, DisasOps *o)
1562 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1563 gen_helper_clgeb(o->out, cpu_env, o->in2, m3);
1564 tcg_temp_free_i32(m3);
1565 gen_set_cc_nz_f32(s, o->in2);
1569 static ExitStatus op_clgdb(DisasContext *s, DisasOps *o)
1571 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1572 gen_helper_clgdb(o->out, cpu_env, o->in2, m3);
1573 tcg_temp_free_i32(m3);
1574 gen_set_cc_nz_f64(s, o->in2);
1578 static ExitStatus op_clgxb(DisasContext *s, DisasOps *o)
1580 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1581 gen_helper_clgxb(o->out, cpu_env, o->in1, o->in2, m3);
1582 tcg_temp_free_i32(m3);
1583 gen_set_cc_nz_f128(s, o->in1, o->in2);
1587 static ExitStatus op_cegb(DisasContext *s, DisasOps *o)
1589 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1590 gen_helper_cegb(o->out, cpu_env, o->in2, m3);
1591 tcg_temp_free_i32(m3);
1595 static ExitStatus op_cdgb(DisasContext *s, DisasOps *o)
1597 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1598 gen_helper_cdgb(o->out, cpu_env, o->in2, m3);
1599 tcg_temp_free_i32(m3);
1603 static ExitStatus op_cxgb(DisasContext *s, DisasOps *o)
1605 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1606 gen_helper_cxgb(o->out, cpu_env, o->in2, m3);
1607 tcg_temp_free_i32(m3);
1608 return_low128(o->out2);
1612 static ExitStatus op_celgb(DisasContext *s, DisasOps *o)
1614 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1615 gen_helper_celgb(o->out, cpu_env, o->in2, m3);
1616 tcg_temp_free_i32(m3);
1620 static ExitStatus op_cdlgb(DisasContext *s, DisasOps *o)
1622 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1623 gen_helper_cdlgb(o->out, cpu_env, o->in2, m3);
1624 tcg_temp_free_i32(m3);
1628 static ExitStatus op_cxlgb(DisasContext *s, DisasOps *o)
1630 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1631 gen_helper_cxlgb(o->out, cpu_env, o->in2, m3);
1632 tcg_temp_free_i32(m3);
1633 return_low128(o->out2);
1637 static ExitStatus op_cksm(DisasContext *s, DisasOps *o)
1639 int r2 = get_field(s->fields, r2);
1640 TCGv_i64 len = tcg_temp_new_i64();
1642 potential_page_fault(s);
1643 gen_helper_cksm(len, cpu_env, o->in1, o->in2, regs[r2 + 1]);
1645 return_low128(o->out);
1647 tcg_gen_add_i64(regs[r2], regs[r2], len);
1648 tcg_gen_sub_i64(regs[r2 + 1], regs[r2 + 1], len);
1649 tcg_temp_free_i64(len);
1654 static ExitStatus op_clc(DisasContext *s, DisasOps *o)
1656 int l = get_field(s->fields, l1);
1661 tcg_gen_qemu_ld8u(cc_src, o->addr1, get_mem_index(s));
1662 tcg_gen_qemu_ld8u(cc_dst, o->in2, get_mem_index(s));
1665 tcg_gen_qemu_ld16u(cc_src, o->addr1, get_mem_index(s));
1666 tcg_gen_qemu_ld16u(cc_dst, o->in2, get_mem_index(s));
1669 tcg_gen_qemu_ld32u(cc_src, o->addr1, get_mem_index(s));
1670 tcg_gen_qemu_ld32u(cc_dst, o->in2, get_mem_index(s));
1673 tcg_gen_qemu_ld64(cc_src, o->addr1, get_mem_index(s));
1674 tcg_gen_qemu_ld64(cc_dst, o->in2, get_mem_index(s));
1677 potential_page_fault(s);
1678 vl = tcg_const_i32(l);
1679 gen_helper_clc(cc_op, cpu_env, vl, o->addr1, o->in2);
1680 tcg_temp_free_i32(vl);
1684 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, cc_src, cc_dst);
1688 static ExitStatus op_clcle(DisasContext *s, DisasOps *o)
1690 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1691 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1692 potential_page_fault(s);
1693 gen_helper_clcle(cc_op, cpu_env, r1, o->in2, r3);
1694 tcg_temp_free_i32(r1);
1695 tcg_temp_free_i32(r3);
1700 static ExitStatus op_clm(DisasContext *s, DisasOps *o)
1702 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1703 TCGv_i32 t1 = tcg_temp_new_i32();
1704 tcg_gen_trunc_i64_i32(t1, o->in1);
1705 potential_page_fault(s);
1706 gen_helper_clm(cc_op, cpu_env, t1, m3, o->in2);
1708 tcg_temp_free_i32(t1);
1709 tcg_temp_free_i32(m3);
1713 static ExitStatus op_clst(DisasContext *s, DisasOps *o)
1715 potential_page_fault(s);
1716 gen_helper_clst(o->in1, cpu_env, regs[0], o->in1, o->in2);
1718 return_low128(o->in2);
1722 static ExitStatus op_cs(DisasContext *s, DisasOps *o)
1724 int r3 = get_field(s->fields, r3);
1725 potential_page_fault(s);
1726 gen_helper_cs(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1731 static ExitStatus op_csg(DisasContext *s, DisasOps *o)
1733 int r3 = get_field(s->fields, r3);
1734 potential_page_fault(s);
1735 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, regs[r3]);
1740 #ifndef CONFIG_USER_ONLY
1741 static ExitStatus op_csp(DisasContext *s, DisasOps *o)
1743 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1744 check_privileged(s);
1745 gen_helper_csp(cc_op, cpu_env, r1, o->in2);
1746 tcg_temp_free_i32(r1);
1752 static ExitStatus op_cds(DisasContext *s, DisasOps *o)
1754 int r3 = get_field(s->fields, r3);
1755 TCGv_i64 in3 = tcg_temp_new_i64();
1756 tcg_gen_deposit_i64(in3, regs[r3 + 1], regs[r3], 32, 32);
1757 potential_page_fault(s);
1758 gen_helper_csg(o->out, cpu_env, o->in1, o->in2, in3);
1759 tcg_temp_free_i64(in3);
1764 static ExitStatus op_cdsg(DisasContext *s, DisasOps *o)
1766 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1767 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1768 potential_page_fault(s);
1769 /* XXX rewrite in tcg */
1770 gen_helper_cdsg(cc_op, cpu_env, r1, o->in2, r3);
1775 static ExitStatus op_cvd(DisasContext *s, DisasOps *o)
1777 TCGv_i64 t1 = tcg_temp_new_i64();
1778 TCGv_i32 t2 = tcg_temp_new_i32();
1779 tcg_gen_trunc_i64_i32(t2, o->in1);
1780 gen_helper_cvd(t1, t2);
1781 tcg_temp_free_i32(t2);
1782 tcg_gen_qemu_st64(t1, o->in2, get_mem_index(s));
1783 tcg_temp_free_i64(t1);
1787 static ExitStatus op_ct(DisasContext *s, DisasOps *o)
1789 int m3 = get_field(s->fields, m3);
1790 int lab = gen_new_label();
1794 /* Bit 3 of the m3 field is reserved and should be zero.
1795 Choose to ignore it wrt the ltgt_cond table above. */
1796 c = tcg_invert_cond(ltgt_cond[m3 & 14]);
1797 if (s->insn->data) {
1798 c = tcg_unsigned_cond(c);
1800 tcg_gen_brcond_i64(c, o->in1, o->in2, lab);
1802 /* Set DXC to 0xff. */
1803 t = tcg_temp_new_i32();
1804 tcg_gen_ld_i32(t, cpu_env, offsetof(CPUS390XState, fpc));
1805 tcg_gen_ori_i32(t, t, 0xff00);
1806 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, fpc));
1807 tcg_temp_free_i32(t);
1810 gen_program_exception(s, PGM_DATA);
1816 #ifndef CONFIG_USER_ONLY
1817 static ExitStatus op_diag(DisasContext *s, DisasOps *o)
1821 check_privileged(s);
1822 potential_page_fault(s);
1824 /* We pretend the format is RX_a so that D2 is the field we want. */
1825 tmp = tcg_const_i32(get_field(s->fields, d2) & 0xfff);
1826 gen_helper_diag(regs[2], cpu_env, tmp, regs[2], regs[1]);
1827 tcg_temp_free_i32(tmp);
1832 static ExitStatus op_divs32(DisasContext *s, DisasOps *o)
1834 gen_helper_divs32(o->out2, cpu_env, o->in1, o->in2);
1835 return_low128(o->out);
1839 static ExitStatus op_divu32(DisasContext *s, DisasOps *o)
1841 gen_helper_divu32(o->out2, cpu_env, o->in1, o->in2);
1842 return_low128(o->out);
1846 static ExitStatus op_divs64(DisasContext *s, DisasOps *o)
1848 gen_helper_divs64(o->out2, cpu_env, o->in1, o->in2);
1849 return_low128(o->out);
1853 static ExitStatus op_divu64(DisasContext *s, DisasOps *o)
1855 gen_helper_divu64(o->out2, cpu_env, o->out, o->out2, o->in2);
1856 return_low128(o->out);
1860 static ExitStatus op_deb(DisasContext *s, DisasOps *o)
1862 gen_helper_deb(o->out, cpu_env, o->in1, o->in2);
1866 static ExitStatus op_ddb(DisasContext *s, DisasOps *o)
1868 gen_helper_ddb(o->out, cpu_env, o->in1, o->in2);
1872 static ExitStatus op_dxb(DisasContext *s, DisasOps *o)
1874 gen_helper_dxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1875 return_low128(o->out2);
1879 static ExitStatus op_ear(DisasContext *s, DisasOps *o)
1881 int r2 = get_field(s->fields, r2);
1882 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, aregs[r2]));
1886 static ExitStatus op_efpc(DisasContext *s, DisasOps *o)
1888 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, fpc));
1892 static ExitStatus op_ex(DisasContext *s, DisasOps *o)
1894 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
1895 tb->flags, (ab)use the tb->cs_base field as the address of
1896 the template in memory, and grab 8 bits of tb->flags/cflags for
1897 the contents of the register. We would then recognize all this
1898 in gen_intermediate_code_internal, generating code for exactly
1899 one instruction. This new TB then gets executed normally.
1901 On the other hand, this seems to be mostly used for modifying
1902 MVC inside of memcpy, which needs a helper call anyway. So
1903 perhaps this doesn't bear thinking about any further. */
1910 tmp = tcg_const_i64(s->next_pc);
1911 gen_helper_ex(cc_op, cpu_env, cc_op, o->in1, o->in2, tmp);
1912 tcg_temp_free_i64(tmp);
1918 static ExitStatus op_flogr(DisasContext *s, DisasOps *o)
1920 /* We'll use the original input for cc computation, since we get to
1921 compare that against 0, which ought to be better than comparing
1922 the real output against 64. It also lets cc_dst be a convenient
1923 temporary during our computation. */
1924 gen_op_update1_cc_i64(s, CC_OP_FLOGR, o->in2);
1926 /* R1 = IN ? CLZ(IN) : 64. */
1927 gen_helper_clz(o->out, o->in2);
1929 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
1930 value by 64, which is undefined. But since the shift is 64 iff the
1931 input is zero, we still get the correct result after and'ing. */
1932 tcg_gen_movi_i64(o->out2, 0x8000000000000000ull);
1933 tcg_gen_shr_i64(o->out2, o->out2, o->out);
1934 tcg_gen_andc_i64(o->out2, cc_dst, o->out2);
1938 static ExitStatus op_icm(DisasContext *s, DisasOps *o)
1940 int m3 = get_field(s->fields, m3);
1941 int pos, len, base = s->insn->data;
1942 TCGv_i64 tmp = tcg_temp_new_i64();
1947 /* Effectively a 32-bit load. */
1948 tcg_gen_qemu_ld32u(tmp, o->in2, get_mem_index(s));
1955 /* Effectively a 16-bit load. */
1956 tcg_gen_qemu_ld16u(tmp, o->in2, get_mem_index(s));
1964 /* Effectively an 8-bit load. */
1965 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
1970 pos = base + ctz32(m3) * 8;
1971 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, len);
1972 ccm = ((1ull << len) - 1) << pos;
1976 /* This is going to be a sequence of loads and inserts. */
1977 pos = base + 32 - 8;
1981 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
1982 tcg_gen_addi_i64(o->in2, o->in2, 1);
1983 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, 8);
1986 m3 = (m3 << 1) & 0xf;
1992 tcg_gen_movi_i64(tmp, ccm);
1993 gen_op_update2_cc_i64(s, CC_OP_ICM, tmp, o->out);
1994 tcg_temp_free_i64(tmp);
1998 static ExitStatus op_insi(DisasContext *s, DisasOps *o)
2000 int shift = s->insn->data & 0xff;
2001 int size = s->insn->data >> 8;
2002 tcg_gen_deposit_i64(o->out, o->in1, o->in2, shift, size);
2006 static ExitStatus op_ipm(DisasContext *s, DisasOps *o)
2011 tcg_gen_andi_i64(o->out, o->out, ~0xff000000ull);
2013 t1 = tcg_temp_new_i64();
2014 tcg_gen_shli_i64(t1, psw_mask, 20);
2015 tcg_gen_shri_i64(t1, t1, 36);
2016 tcg_gen_or_i64(o->out, o->out, t1);
2018 tcg_gen_extu_i32_i64(t1, cc_op);
2019 tcg_gen_shli_i64(t1, t1, 28);
2020 tcg_gen_or_i64(o->out, o->out, t1);
2021 tcg_temp_free_i64(t1);
2025 #ifndef CONFIG_USER_ONLY
2026 static ExitStatus op_ipte(DisasContext *s, DisasOps *o)
2028 check_privileged(s);
2029 gen_helper_ipte(cpu_env, o->in1, o->in2);
2033 static ExitStatus op_iske(DisasContext *s, DisasOps *o)
2035 check_privileged(s);
2036 gen_helper_iske(o->out, cpu_env, o->in2);
2041 static ExitStatus op_ldeb(DisasContext *s, DisasOps *o)
2043 gen_helper_ldeb(o->out, cpu_env, o->in2);
2047 static ExitStatus op_ledb(DisasContext *s, DisasOps *o)
2049 gen_helper_ledb(o->out, cpu_env, o->in2);
2053 static ExitStatus op_ldxb(DisasContext *s, DisasOps *o)
2055 gen_helper_ldxb(o->out, cpu_env, o->in1, o->in2);
2059 static ExitStatus op_lexb(DisasContext *s, DisasOps *o)
2061 gen_helper_lexb(o->out, cpu_env, o->in1, o->in2);
2065 static ExitStatus op_lxdb(DisasContext *s, DisasOps *o)
2067 gen_helper_lxdb(o->out, cpu_env, o->in2);
2068 return_low128(o->out2);
2072 static ExitStatus op_lxeb(DisasContext *s, DisasOps *o)
2074 gen_helper_lxeb(o->out, cpu_env, o->in2);
2075 return_low128(o->out2);
2079 static ExitStatus op_llgt(DisasContext *s, DisasOps *o)
2081 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffff);
2085 static ExitStatus op_ld8s(DisasContext *s, DisasOps *o)
2087 tcg_gen_qemu_ld8s(o->out, o->in2, get_mem_index(s));
2091 static ExitStatus op_ld8u(DisasContext *s, DisasOps *o)
2093 tcg_gen_qemu_ld8u(o->out, o->in2, get_mem_index(s));
2097 static ExitStatus op_ld16s(DisasContext *s, DisasOps *o)
2099 tcg_gen_qemu_ld16s(o->out, o->in2, get_mem_index(s));
2103 static ExitStatus op_ld16u(DisasContext *s, DisasOps *o)
2105 tcg_gen_qemu_ld16u(o->out, o->in2, get_mem_index(s));
2109 static ExitStatus op_ld32s(DisasContext *s, DisasOps *o)
2111 tcg_gen_qemu_ld32s(o->out, o->in2, get_mem_index(s));
2115 static ExitStatus op_ld32u(DisasContext *s, DisasOps *o)
2117 tcg_gen_qemu_ld32u(o->out, o->in2, get_mem_index(s));
2121 static ExitStatus op_ld64(DisasContext *s, DisasOps *o)
2123 tcg_gen_qemu_ld64(o->out, o->in2, get_mem_index(s));
2127 static ExitStatus op_loc(DisasContext *s, DisasOps *o)
2131 disas_jcc(s, &c, get_field(s->fields, m3));
2134 tcg_gen_movcond_i64(c.cond, o->out, c.u.s64.a, c.u.s64.b,
2138 TCGv_i32 t32 = tcg_temp_new_i32();
2141 tcg_gen_setcond_i32(c.cond, t32, c.u.s32.a, c.u.s32.b);
2144 t = tcg_temp_new_i64();
2145 tcg_gen_extu_i32_i64(t, t32);
2146 tcg_temp_free_i32(t32);
2148 z = tcg_const_i64(0);
2149 tcg_gen_movcond_i64(TCG_COND_NE, o->out, t, z, o->in2, o->in1);
2150 tcg_temp_free_i64(t);
2151 tcg_temp_free_i64(z);
2157 #ifndef CONFIG_USER_ONLY
2158 static ExitStatus op_lctl(DisasContext *s, DisasOps *o)
2160 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2161 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2162 check_privileged(s);
2163 potential_page_fault(s);
2164 gen_helper_lctl(cpu_env, r1, o->in2, r3);
2165 tcg_temp_free_i32(r1);
2166 tcg_temp_free_i32(r3);
2170 static ExitStatus op_lctlg(DisasContext *s, DisasOps *o)
2172 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2173 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2174 check_privileged(s);
2175 potential_page_fault(s);
2176 gen_helper_lctlg(cpu_env, r1, o->in2, r3);
2177 tcg_temp_free_i32(r1);
2178 tcg_temp_free_i32(r3);
2181 static ExitStatus op_lra(DisasContext *s, DisasOps *o)
2183 check_privileged(s);
2184 potential_page_fault(s);
2185 gen_helper_lra(o->out, cpu_env, o->in2);
2190 static ExitStatus op_lpsw(DisasContext *s, DisasOps *o)
2194 check_privileged(s);
2196 t1 = tcg_temp_new_i64();
2197 t2 = tcg_temp_new_i64();
2198 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2199 tcg_gen_addi_i64(o->in2, o->in2, 4);
2200 tcg_gen_qemu_ld32u(t2, o->in2, get_mem_index(s));
2201 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2202 tcg_gen_shli_i64(t1, t1, 32);
2203 gen_helper_load_psw(cpu_env, t1, t2);
2204 tcg_temp_free_i64(t1);
2205 tcg_temp_free_i64(t2);
2206 return EXIT_NORETURN;
2209 static ExitStatus op_lpswe(DisasContext *s, DisasOps *o)
2213 check_privileged(s);
2215 t1 = tcg_temp_new_i64();
2216 t2 = tcg_temp_new_i64();
2217 tcg_gen_qemu_ld64(t1, o->in2, get_mem_index(s));
2218 tcg_gen_addi_i64(o->in2, o->in2, 8);
2219 tcg_gen_qemu_ld64(t2, o->in2, get_mem_index(s));
2220 gen_helper_load_psw(cpu_env, t1, t2);
2221 tcg_temp_free_i64(t1);
2222 tcg_temp_free_i64(t2);
2223 return EXIT_NORETURN;
2227 static ExitStatus op_lam(DisasContext *s, DisasOps *o)
2229 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2230 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2231 potential_page_fault(s);
2232 gen_helper_lam(cpu_env, r1, o->in2, r3);
2233 tcg_temp_free_i32(r1);
2234 tcg_temp_free_i32(r3);
2238 static ExitStatus op_lm32(DisasContext *s, DisasOps *o)
2240 int r1 = get_field(s->fields, r1);
2241 int r3 = get_field(s->fields, r3);
2242 TCGv_i64 t = tcg_temp_new_i64();
2243 TCGv_i64 t4 = tcg_const_i64(4);
2246 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2247 store_reg32_i64(r1, t);
2251 tcg_gen_add_i64(o->in2, o->in2, t4);
2255 tcg_temp_free_i64(t);
2256 tcg_temp_free_i64(t4);
2260 static ExitStatus op_lmh(DisasContext *s, DisasOps *o)
2262 int r1 = get_field(s->fields, r1);
2263 int r3 = get_field(s->fields, r3);
2264 TCGv_i64 t = tcg_temp_new_i64();
2265 TCGv_i64 t4 = tcg_const_i64(4);
2268 tcg_gen_qemu_ld32u(t, o->in2, get_mem_index(s));
2269 store_reg32h_i64(r1, t);
2273 tcg_gen_add_i64(o->in2, o->in2, t4);
2277 tcg_temp_free_i64(t);
2278 tcg_temp_free_i64(t4);
2282 static ExitStatus op_lm64(DisasContext *s, DisasOps *o)
2284 int r1 = get_field(s->fields, r1);
2285 int r3 = get_field(s->fields, r3);
2286 TCGv_i64 t8 = tcg_const_i64(8);
2289 tcg_gen_qemu_ld64(regs[r1], o->in2, get_mem_index(s));
2293 tcg_gen_add_i64(o->in2, o->in2, t8);
2297 tcg_temp_free_i64(t8);
2301 static ExitStatus op_mov2(DisasContext *s, DisasOps *o)
2304 o->g_out = o->g_in2;
2305 TCGV_UNUSED_I64(o->in2);
2310 static ExitStatus op_movx(DisasContext *s, DisasOps *o)
2314 o->g_out = o->g_in1;
2315 o->g_out2 = o->g_in2;
2316 TCGV_UNUSED_I64(o->in1);
2317 TCGV_UNUSED_I64(o->in2);
2318 o->g_in1 = o->g_in2 = false;
2322 static ExitStatus op_mvc(DisasContext *s, DisasOps *o)
2324 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2325 potential_page_fault(s);
2326 gen_helper_mvc(cpu_env, l, o->addr1, o->in2);
2327 tcg_temp_free_i32(l);
2331 static ExitStatus op_mvcl(DisasContext *s, DisasOps *o)
2333 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2334 TCGv_i32 r2 = tcg_const_i32(get_field(s->fields, r2));
2335 potential_page_fault(s);
2336 gen_helper_mvcl(cc_op, cpu_env, r1, r2);
2337 tcg_temp_free_i32(r1);
2338 tcg_temp_free_i32(r2);
2343 static ExitStatus op_mvcle(DisasContext *s, DisasOps *o)
2345 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2346 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2347 potential_page_fault(s);
2348 gen_helper_mvcle(cc_op, cpu_env, r1, o->in2, r3);
2349 tcg_temp_free_i32(r1);
2350 tcg_temp_free_i32(r3);
2355 #ifndef CONFIG_USER_ONLY
2356 static ExitStatus op_mvcp(DisasContext *s, DisasOps *o)
2358 int r1 = get_field(s->fields, l1);
2359 check_privileged(s);
2360 potential_page_fault(s);
2361 gen_helper_mvcp(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2366 static ExitStatus op_mvcs(DisasContext *s, DisasOps *o)
2368 int r1 = get_field(s->fields, l1);
2369 check_privileged(s);
2370 potential_page_fault(s);
2371 gen_helper_mvcs(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2377 static ExitStatus op_mvpg(DisasContext *s, DisasOps *o)
2379 potential_page_fault(s);
2380 gen_helper_mvpg(cpu_env, regs[0], o->in1, o->in2);
2385 static ExitStatus op_mvst(DisasContext *s, DisasOps *o)
2387 potential_page_fault(s);
2388 gen_helper_mvst(o->in1, cpu_env, regs[0], o->in1, o->in2);
2390 return_low128(o->in2);
2394 static ExitStatus op_mul(DisasContext *s, DisasOps *o)
2396 tcg_gen_mul_i64(o->out, o->in1, o->in2);
2400 static ExitStatus op_mul128(DisasContext *s, DisasOps *o)
2402 gen_helper_mul128(o->out, cpu_env, o->in1, o->in2);
2403 return_low128(o->out2);
2407 static ExitStatus op_meeb(DisasContext *s, DisasOps *o)
2409 gen_helper_meeb(o->out, cpu_env, o->in1, o->in2);
2413 static ExitStatus op_mdeb(DisasContext *s, DisasOps *o)
2415 gen_helper_mdeb(o->out, cpu_env, o->in1, o->in2);
2419 static ExitStatus op_mdb(DisasContext *s, DisasOps *o)
2421 gen_helper_mdb(o->out, cpu_env, o->in1, o->in2);
2425 static ExitStatus op_mxb(DisasContext *s, DisasOps *o)
2427 gen_helper_mxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2428 return_low128(o->out2);
2432 static ExitStatus op_mxdb(DisasContext *s, DisasOps *o)
2434 gen_helper_mxdb(o->out, cpu_env, o->out, o->out2, o->in2);
2435 return_low128(o->out2);
2439 static ExitStatus op_maeb(DisasContext *s, DisasOps *o)
2441 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2442 gen_helper_maeb(o->out, cpu_env, o->in1, o->in2, r3);
2443 tcg_temp_free_i64(r3);
2447 static ExitStatus op_madb(DisasContext *s, DisasOps *o)
2449 int r3 = get_field(s->fields, r3);
2450 gen_helper_madb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2454 static ExitStatus op_mseb(DisasContext *s, DisasOps *o)
2456 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2457 gen_helper_mseb(o->out, cpu_env, o->in1, o->in2, r3);
2458 tcg_temp_free_i64(r3);
2462 static ExitStatus op_msdb(DisasContext *s, DisasOps *o)
2464 int r3 = get_field(s->fields, r3);
2465 gen_helper_msdb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2469 static ExitStatus op_nabs(DisasContext *s, DisasOps *o)
2471 gen_helper_nabs_i64(o->out, o->in2);
2475 static ExitStatus op_nabsf32(DisasContext *s, DisasOps *o)
2477 tcg_gen_ori_i64(o->out, o->in2, 0x80000000ull);
2481 static ExitStatus op_nabsf64(DisasContext *s, DisasOps *o)
2483 tcg_gen_ori_i64(o->out, o->in2, 0x8000000000000000ull);
2487 static ExitStatus op_nabsf128(DisasContext *s, DisasOps *o)
2489 tcg_gen_ori_i64(o->out, o->in1, 0x8000000000000000ull);
2490 tcg_gen_mov_i64(o->out2, o->in2);
2494 static ExitStatus op_nc(DisasContext *s, DisasOps *o)
2496 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2497 potential_page_fault(s);
2498 gen_helper_nc(cc_op, cpu_env, l, o->addr1, o->in2);
2499 tcg_temp_free_i32(l);
2504 static ExitStatus op_neg(DisasContext *s, DisasOps *o)
2506 tcg_gen_neg_i64(o->out, o->in2);
2510 static ExitStatus op_negf32(DisasContext *s, DisasOps *o)
2512 tcg_gen_xori_i64(o->out, o->in2, 0x80000000ull);
2516 static ExitStatus op_negf64(DisasContext *s, DisasOps *o)
2518 tcg_gen_xori_i64(o->out, o->in2, 0x8000000000000000ull);
2522 static ExitStatus op_negf128(DisasContext *s, DisasOps *o)
2524 tcg_gen_xori_i64(o->out, o->in1, 0x8000000000000000ull);
2525 tcg_gen_mov_i64(o->out2, o->in2);
2529 static ExitStatus op_oc(DisasContext *s, DisasOps *o)
2531 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2532 potential_page_fault(s);
2533 gen_helper_oc(cc_op, cpu_env, l, o->addr1, o->in2);
2534 tcg_temp_free_i32(l);
2539 static ExitStatus op_or(DisasContext *s, DisasOps *o)
2541 tcg_gen_or_i64(o->out, o->in1, o->in2);
2545 static ExitStatus op_ori(DisasContext *s, DisasOps *o)
2547 int shift = s->insn->data & 0xff;
2548 int size = s->insn->data >> 8;
2549 uint64_t mask = ((1ull << size) - 1) << shift;
2552 tcg_gen_shli_i64(o->in2, o->in2, shift);
2553 tcg_gen_or_i64(o->out, o->in1, o->in2);
2555 /* Produce the CC from only the bits manipulated. */
2556 tcg_gen_andi_i64(cc_dst, o->out, mask);
2557 set_cc_nz_u64(s, cc_dst);
2561 static ExitStatus op_popcnt(DisasContext *s, DisasOps *o)
2563 gen_helper_popcnt(o->out, o->in2);
2567 #ifndef CONFIG_USER_ONLY
2568 static ExitStatus op_ptlb(DisasContext *s, DisasOps *o)
2570 check_privileged(s);
2571 gen_helper_ptlb(cpu_env);
2576 static ExitStatus op_risbg(DisasContext *s, DisasOps *o)
2578 int i3 = get_field(s->fields, i3);
2579 int i4 = get_field(s->fields, i4);
2580 int i5 = get_field(s->fields, i5);
2581 int do_zero = i4 & 0x80;
2582 uint64_t mask, imask, pmask;
2585 /* Adjust the arguments for the specific insn. */
2586 switch (s->fields->op2) {
2587 case 0x55: /* risbg */
2592 case 0x5d: /* risbhg */
2595 pmask = 0xffffffff00000000ull;
2597 case 0x51: /* risblg */
2600 pmask = 0x00000000ffffffffull;
2606 /* MASK is the set of bits to be inserted from R2.
2607 Take care for I3/I4 wraparound. */
2610 mask ^= pmask >> i4 >> 1;
2612 mask |= ~(pmask >> i4 >> 1);
2616 /* IMASK is the set of bits to be kept from R1. In the case of the high/low
2617 insns, we need to keep the other half of the register. */
2618 imask = ~mask | ~pmask;
2620 if (s->fields->op2 == 0x55) {
2627 /* In some cases we can implement this with deposit, which can be more
2628 efficient on some hosts. */
2629 if (~mask == imask && i3 <= i4) {
2630 if (s->fields->op2 == 0x5d) {
2633 /* Note that we rotate the bits to be inserted to the lsb, not to
2634 the position as described in the PoO. */
2637 rot = (i5 - pos) & 63;
2643 /* Rotate the input as necessary. */
2644 tcg_gen_rotli_i64(o->in2, o->in2, rot);
2646 /* Insert the selected bits into the output. */
2648 tcg_gen_deposit_i64(o->out, o->out, o->in2, pos, len);
2649 } else if (imask == 0) {
2650 tcg_gen_andi_i64(o->out, o->in2, mask);
2652 tcg_gen_andi_i64(o->in2, o->in2, mask);
2653 tcg_gen_andi_i64(o->out, o->out, imask);
2654 tcg_gen_or_i64(o->out, o->out, o->in2);
2659 static ExitStatus op_rosbg(DisasContext *s, DisasOps *o)
2661 int i3 = get_field(s->fields, i3);
2662 int i4 = get_field(s->fields, i4);
2663 int i5 = get_field(s->fields, i5);
2666 /* If this is a test-only form, arrange to discard the result. */
2668 o->out = tcg_temp_new_i64();
2676 /* MASK is the set of bits to be operated on from R2.
2677 Take care for I3/I4 wraparound. */
2680 mask ^= ~0ull >> i4 >> 1;
2682 mask |= ~(~0ull >> i4 >> 1);
2685 /* Rotate the input as necessary. */
2686 tcg_gen_rotli_i64(o->in2, o->in2, i5);
2689 switch (s->fields->op2) {
2690 case 0x55: /* AND */
2691 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
2692 tcg_gen_and_i64(o->out, o->out, o->in2);
2695 tcg_gen_andi_i64(o->in2, o->in2, mask);
2696 tcg_gen_or_i64(o->out, o->out, o->in2);
2698 case 0x57: /* XOR */
2699 tcg_gen_andi_i64(o->in2, o->in2, mask);
2700 tcg_gen_xor_i64(o->out, o->out, o->in2);
2707 tcg_gen_andi_i64(cc_dst, o->out, mask);
2708 set_cc_nz_u64(s, cc_dst);
2712 static ExitStatus op_rev16(DisasContext *s, DisasOps *o)
2714 tcg_gen_bswap16_i64(o->out, o->in2);
2718 static ExitStatus op_rev32(DisasContext *s, DisasOps *o)
2720 tcg_gen_bswap32_i64(o->out, o->in2);
2724 static ExitStatus op_rev64(DisasContext *s, DisasOps *o)
2726 tcg_gen_bswap64_i64(o->out, o->in2);
2730 static ExitStatus op_rll32(DisasContext *s, DisasOps *o)
2732 TCGv_i32 t1 = tcg_temp_new_i32();
2733 TCGv_i32 t2 = tcg_temp_new_i32();
2734 TCGv_i32 to = tcg_temp_new_i32();
2735 tcg_gen_trunc_i64_i32(t1, o->in1);
2736 tcg_gen_trunc_i64_i32(t2, o->in2);
2737 tcg_gen_rotl_i32(to, t1, t2);
2738 tcg_gen_extu_i32_i64(o->out, to);
2739 tcg_temp_free_i32(t1);
2740 tcg_temp_free_i32(t2);
2741 tcg_temp_free_i32(to);
2745 static ExitStatus op_rll64(DisasContext *s, DisasOps *o)
2747 tcg_gen_rotl_i64(o->out, o->in1, o->in2);
2751 #ifndef CONFIG_USER_ONLY
2752 static ExitStatus op_rrbe(DisasContext *s, DisasOps *o)
2754 check_privileged(s);
2755 gen_helper_rrbe(cc_op, cpu_env, o->in2);
2760 static ExitStatus op_sacf(DisasContext *s, DisasOps *o)
2762 check_privileged(s);
2763 gen_helper_sacf(cpu_env, o->in2);
2764 /* Addressing mode has changed, so end the block. */
2765 return EXIT_PC_STALE;
2769 static ExitStatus op_sar(DisasContext *s, DisasOps *o)
2771 int r1 = get_field(s->fields, r1);
2772 tcg_gen_st32_i64(o->in2, cpu_env, offsetof(CPUS390XState, aregs[r1]));
2776 static ExitStatus op_seb(DisasContext *s, DisasOps *o)
2778 gen_helper_seb(o->out, cpu_env, o->in1, o->in2);
2782 static ExitStatus op_sdb(DisasContext *s, DisasOps *o)
2784 gen_helper_sdb(o->out, cpu_env, o->in1, o->in2);
2788 static ExitStatus op_sxb(DisasContext *s, DisasOps *o)
2790 gen_helper_sxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2791 return_low128(o->out2);
2795 static ExitStatus op_sqeb(DisasContext *s, DisasOps *o)
2797 gen_helper_sqeb(o->out, cpu_env, o->in2);
2801 static ExitStatus op_sqdb(DisasContext *s, DisasOps *o)
2803 gen_helper_sqdb(o->out, cpu_env, o->in2);
2807 static ExitStatus op_sqxb(DisasContext *s, DisasOps *o)
2809 gen_helper_sqxb(o->out, cpu_env, o->in1, o->in2);
2810 return_low128(o->out2);
2814 #ifndef CONFIG_USER_ONLY
2815 static ExitStatus op_servc(DisasContext *s, DisasOps *o)
2817 check_privileged(s);
2818 potential_page_fault(s);
2819 gen_helper_servc(cc_op, cpu_env, o->in2, o->in1);
2824 static ExitStatus op_sigp(DisasContext *s, DisasOps *o)
2826 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2827 check_privileged(s);
2828 potential_page_fault(s);
2829 gen_helper_sigp(cc_op, cpu_env, o->in2, r1, o->in1);
2830 tcg_temp_free_i32(r1);
2835 static ExitStatus op_soc(DisasContext *s, DisasOps *o)
2841 disas_jcc(s, &c, get_field(s->fields, m3));
2843 lab = gen_new_label();
2845 tcg_gen_brcond_i64(c.cond, c.u.s64.a, c.u.s64.b, lab);
2847 tcg_gen_brcond_i32(c.cond, c.u.s32.a, c.u.s32.b, lab);
2851 r1 = get_field(s->fields, r1);
2852 a = get_address(s, 0, get_field(s->fields, b2), get_field(s->fields, d2));
2853 if (s->insn->data) {
2854 tcg_gen_qemu_st64(regs[r1], a, get_mem_index(s));
2856 tcg_gen_qemu_st32(regs[r1], a, get_mem_index(s));
2858 tcg_temp_free_i64(a);
2864 static ExitStatus op_sla(DisasContext *s, DisasOps *o)
2866 uint64_t sign = 1ull << s->insn->data;
2867 enum cc_op cco = s->insn->data == 31 ? CC_OP_SLA_32 : CC_OP_SLA_64;
2868 gen_op_update2_cc_i64(s, cco, o->in1, o->in2);
2869 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2870 /* The arithmetic left shift is curious in that it does not affect
2871 the sign bit. Copy that over from the source unchanged. */
2872 tcg_gen_andi_i64(o->out, o->out, ~sign);
2873 tcg_gen_andi_i64(o->in1, o->in1, sign);
2874 tcg_gen_or_i64(o->out, o->out, o->in1);
2878 static ExitStatus op_sll(DisasContext *s, DisasOps *o)
2880 tcg_gen_shl_i64(o->out, o->in1, o->in2);
2884 static ExitStatus op_sra(DisasContext *s, DisasOps *o)
2886 tcg_gen_sar_i64(o->out, o->in1, o->in2);
2890 static ExitStatus op_srl(DisasContext *s, DisasOps *o)
2892 tcg_gen_shr_i64(o->out, o->in1, o->in2);
2896 static ExitStatus op_sfpc(DisasContext *s, DisasOps *o)
2898 gen_helper_sfpc(cpu_env, o->in2);
2902 #ifndef CONFIG_USER_ONLY
2903 static ExitStatus op_spka(DisasContext *s, DisasOps *o)
2905 check_privileged(s);
2906 tcg_gen_shri_i64(o->in2, o->in2, 4);
2907 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, PSW_SHIFT_KEY - 4, 4);
2911 static ExitStatus op_sske(DisasContext *s, DisasOps *o)
2913 check_privileged(s);
2914 gen_helper_sske(cpu_env, o->in1, o->in2);
2918 static ExitStatus op_ssm(DisasContext *s, DisasOps *o)
2920 check_privileged(s);
2921 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, 56, 8);
2925 static ExitStatus op_stap(DisasContext *s, DisasOps *o)
2927 check_privileged(s);
2928 /* ??? Surely cpu address != cpu number. In any case the previous
2929 version of this stored more than the required half-word, so it
2930 is unlikely this has ever been tested. */
2931 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
2935 static ExitStatus op_stck(DisasContext *s, DisasOps *o)
2937 gen_helper_stck(o->out, cpu_env);
2938 /* ??? We don't implement clock states. */
2939 gen_op_movi_cc(s, 0);
2943 static ExitStatus op_stcke(DisasContext *s, DisasOps *o)
2945 TCGv_i64 c1 = tcg_temp_new_i64();
2946 TCGv_i64 c2 = tcg_temp_new_i64();
2947 gen_helper_stck(c1, cpu_env);
2948 /* Shift the 64-bit value into its place as a zero-extended
2949 104-bit value. Note that "bit positions 64-103 are always
2950 non-zero so that they compare differently to STCK"; we set
2951 the least significant bit to 1. */
2952 tcg_gen_shli_i64(c2, c1, 56);
2953 tcg_gen_shri_i64(c1, c1, 8);
2954 tcg_gen_ori_i64(c2, c2, 0x10000);
2955 tcg_gen_qemu_st64(c1, o->in2, get_mem_index(s));
2956 tcg_gen_addi_i64(o->in2, o->in2, 8);
2957 tcg_gen_qemu_st64(c2, o->in2, get_mem_index(s));
2958 tcg_temp_free_i64(c1);
2959 tcg_temp_free_i64(c2);
2960 /* ??? We don't implement clock states. */
2961 gen_op_movi_cc(s, 0);
2965 static ExitStatus op_sckc(DisasContext *s, DisasOps *o)
2967 check_privileged(s);
2968 gen_helper_sckc(cpu_env, o->in2);
2972 static ExitStatus op_stckc(DisasContext *s, DisasOps *o)
2974 check_privileged(s);
2975 gen_helper_stckc(o->out, cpu_env);
2979 static ExitStatus op_stctg(DisasContext *s, DisasOps *o)
2981 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2982 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2983 check_privileged(s);
2984 potential_page_fault(s);
2985 gen_helper_stctg(cpu_env, r1, o->in2, r3);
2986 tcg_temp_free_i32(r1);
2987 tcg_temp_free_i32(r3);
2991 static ExitStatus op_stctl(DisasContext *s, DisasOps *o)
2993 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2994 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2995 check_privileged(s);
2996 potential_page_fault(s);
2997 gen_helper_stctl(cpu_env, r1, o->in2, r3);
2998 tcg_temp_free_i32(r1);
2999 tcg_temp_free_i32(r3);
3003 static ExitStatus op_stidp(DisasContext *s, DisasOps *o)
3005 check_privileged(s);
3006 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
3010 static ExitStatus op_spt(DisasContext *s, DisasOps *o)
3012 check_privileged(s);
3013 gen_helper_spt(cpu_env, o->in2);
3017 static ExitStatus op_stfl(DisasContext *s, DisasOps *o)
3020 /* We really ought to have more complete indication of facilities
3021 that we implement. Address this when STFLE is implemented. */
3022 check_privileged(s);
3023 f = tcg_const_i64(0xc0000000);
3024 a = tcg_const_i64(200);
3025 tcg_gen_qemu_st32(f, a, get_mem_index(s));
3026 tcg_temp_free_i64(f);
3027 tcg_temp_free_i64(a);
3031 static ExitStatus op_stpt(DisasContext *s, DisasOps *o)
3033 check_privileged(s);
3034 gen_helper_stpt(o->out, cpu_env);
3038 static ExitStatus op_stsi(DisasContext *s, DisasOps *o)
3040 check_privileged(s);
3041 potential_page_fault(s);
3042 gen_helper_stsi(cc_op, cpu_env, o->in2, regs[0], regs[1]);
3047 static ExitStatus op_spx(DisasContext *s, DisasOps *o)
3049 check_privileged(s);
3050 gen_helper_spx(cpu_env, o->in2);
3054 static ExitStatus op_subchannel(DisasContext *s, DisasOps *o)
3056 check_privileged(s);
3057 /* Not operational. */
3058 gen_op_movi_cc(s, 3);
3062 static ExitStatus op_stpx(DisasContext *s, DisasOps *o)
3064 check_privileged(s);
3065 tcg_gen_ld_i64(o->out, cpu_env, offsetof(CPUS390XState, psa));
3066 tcg_gen_andi_i64(o->out, o->out, 0x7fffe000);
3070 static ExitStatus op_stnosm(DisasContext *s, DisasOps *o)
3072 uint64_t i2 = get_field(s->fields, i2);
3075 check_privileged(s);
3077 /* It is important to do what the instruction name says: STORE THEN.
3078 If we let the output hook perform the store then if we fault and
3079 restart, we'll have the wrong SYSTEM MASK in place. */
3080 t = tcg_temp_new_i64();
3081 tcg_gen_shri_i64(t, psw_mask, 56);
3082 tcg_gen_qemu_st8(t, o->addr1, get_mem_index(s));
3083 tcg_temp_free_i64(t);
3085 if (s->fields->op == 0xac) {
3086 tcg_gen_andi_i64(psw_mask, psw_mask,
3087 (i2 << 56) | 0x00ffffffffffffffull);
3089 tcg_gen_ori_i64(psw_mask, psw_mask, i2 << 56);
3094 static ExitStatus op_stura(DisasContext *s, DisasOps *o)
3096 check_privileged(s);
3097 potential_page_fault(s);
3098 gen_helper_stura(cpu_env, o->in2, o->in1);
3103 static ExitStatus op_st8(DisasContext *s, DisasOps *o)
3105 tcg_gen_qemu_st8(o->in1, o->in2, get_mem_index(s));
3109 static ExitStatus op_st16(DisasContext *s, DisasOps *o)
3111 tcg_gen_qemu_st16(o->in1, o->in2, get_mem_index(s));
3115 static ExitStatus op_st32(DisasContext *s, DisasOps *o)
3117 tcg_gen_qemu_st32(o->in1, o->in2, get_mem_index(s));
3121 static ExitStatus op_st64(DisasContext *s, DisasOps *o)
3123 tcg_gen_qemu_st64(o->in1, o->in2, get_mem_index(s));
3127 static ExitStatus op_stam(DisasContext *s, DisasOps *o)
3129 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
3130 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
3131 potential_page_fault(s);
3132 gen_helper_stam(cpu_env, r1, o->in2, r3);
3133 tcg_temp_free_i32(r1);
3134 tcg_temp_free_i32(r3);
3138 static ExitStatus op_stcm(DisasContext *s, DisasOps *o)
3140 int m3 = get_field(s->fields, m3);
3141 int pos, base = s->insn->data;
3142 TCGv_i64 tmp = tcg_temp_new_i64();
3144 pos = base + ctz32(m3) * 8;
3147 /* Effectively a 32-bit store. */
3148 tcg_gen_shri_i64(tmp, o->in1, pos);
3149 tcg_gen_qemu_st32(tmp, o->in2, get_mem_index(s));
3155 /* Effectively a 16-bit store. */
3156 tcg_gen_shri_i64(tmp, o->in1, pos);
3157 tcg_gen_qemu_st16(tmp, o->in2, get_mem_index(s));
3164 /* Effectively an 8-bit store. */
3165 tcg_gen_shri_i64(tmp, o->in1, pos);
3166 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3170 /* This is going to be a sequence of shifts and stores. */
3171 pos = base + 32 - 8;
3174 tcg_gen_shri_i64(tmp, o->in1, pos);
3175 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3176 tcg_gen_addi_i64(o->in2, o->in2, 1);
3178 m3 = (m3 << 1) & 0xf;
3183 tcg_temp_free_i64(tmp);
3187 static ExitStatus op_stm(DisasContext *s, DisasOps *o)
3189 int r1 = get_field(s->fields, r1);
3190 int r3 = get_field(s->fields, r3);
3191 int size = s->insn->data;
3192 TCGv_i64 tsize = tcg_const_i64(size);
3196 tcg_gen_qemu_st64(regs[r1], o->in2, get_mem_index(s));
3198 tcg_gen_qemu_st32(regs[r1], o->in2, get_mem_index(s));
3203 tcg_gen_add_i64(o->in2, o->in2, tsize);
3207 tcg_temp_free_i64(tsize);
3211 static ExitStatus op_stmh(DisasContext *s, DisasOps *o)
3213 int r1 = get_field(s->fields, r1);
3214 int r3 = get_field(s->fields, r3);
3215 TCGv_i64 t = tcg_temp_new_i64();
3216 TCGv_i64 t4 = tcg_const_i64(4);
3217 TCGv_i64 t32 = tcg_const_i64(32);
3220 tcg_gen_shl_i64(t, regs[r1], t32);
3221 tcg_gen_qemu_st32(t, o->in2, get_mem_index(s));
3225 tcg_gen_add_i64(o->in2, o->in2, t4);
3229 tcg_temp_free_i64(t);
3230 tcg_temp_free_i64(t4);
3231 tcg_temp_free_i64(t32);
3235 static ExitStatus op_srst(DisasContext *s, DisasOps *o)
3237 potential_page_fault(s);
3238 gen_helper_srst(o->in1, cpu_env, regs[0], o->in1, o->in2);
3240 return_low128(o->in2);
3244 static ExitStatus op_sub(DisasContext *s, DisasOps *o)
3246 tcg_gen_sub_i64(o->out, o->in1, o->in2);
3250 static ExitStatus op_subb(DisasContext *s, DisasOps *o)
3255 tcg_gen_not_i64(o->in2, o->in2);
3256 tcg_gen_add_i64(o->out, o->in1, o->in2);
3258 /* XXX possible optimization point */
3260 cc = tcg_temp_new_i64();
3261 tcg_gen_extu_i32_i64(cc, cc_op);
3262 tcg_gen_shri_i64(cc, cc, 1);
3263 tcg_gen_add_i64(o->out, o->out, cc);
3264 tcg_temp_free_i64(cc);
3268 static ExitStatus op_svc(DisasContext *s, DisasOps *o)
3275 t = tcg_const_i32(get_field(s->fields, i1) & 0xff);
3276 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_code));
3277 tcg_temp_free_i32(t);
3279 t = tcg_const_i32(s->next_pc - s->pc);
3280 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_ilen));
3281 tcg_temp_free_i32(t);
3283 gen_exception(EXCP_SVC);
3284 return EXIT_NORETURN;
3287 static ExitStatus op_tceb(DisasContext *s, DisasOps *o)
3289 gen_helper_tceb(cc_op, o->in1, o->in2);
3294 static ExitStatus op_tcdb(DisasContext *s, DisasOps *o)
3296 gen_helper_tcdb(cc_op, o->in1, o->in2);
3301 static ExitStatus op_tcxb(DisasContext *s, DisasOps *o)
3303 gen_helper_tcxb(cc_op, o->out, o->out2, o->in2);
3308 #ifndef CONFIG_USER_ONLY
3309 static ExitStatus op_tprot(DisasContext *s, DisasOps *o)
3311 potential_page_fault(s);
3312 gen_helper_tprot(cc_op, o->addr1, o->in2);
3318 static ExitStatus op_tr(DisasContext *s, DisasOps *o)
3320 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3321 potential_page_fault(s);
3322 gen_helper_tr(cpu_env, l, o->addr1, o->in2);
3323 tcg_temp_free_i32(l);
3328 static ExitStatus op_unpk(DisasContext *s, DisasOps *o)
3330 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3331 potential_page_fault(s);
3332 gen_helper_unpk(cpu_env, l, o->addr1, o->in2);
3333 tcg_temp_free_i32(l);
3337 static ExitStatus op_xc(DisasContext *s, DisasOps *o)
3339 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3340 potential_page_fault(s);
3341 gen_helper_xc(cc_op, cpu_env, l, o->addr1, o->in2);
3342 tcg_temp_free_i32(l);
3347 static ExitStatus op_xor(DisasContext *s, DisasOps *o)
3349 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3353 static ExitStatus op_xori(DisasContext *s, DisasOps *o)
3355 int shift = s->insn->data & 0xff;
3356 int size = s->insn->data >> 8;
3357 uint64_t mask = ((1ull << size) - 1) << shift;
3360 tcg_gen_shli_i64(o->in2, o->in2, shift);
3361 tcg_gen_xor_i64(o->out, o->in1, o->in2);
3363 /* Produce the CC from only the bits manipulated. */
3364 tcg_gen_andi_i64(cc_dst, o->out, mask);
3365 set_cc_nz_u64(s, cc_dst);
3369 static ExitStatus op_zero(DisasContext *s, DisasOps *o)
3371 o->out = tcg_const_i64(0);
3375 static ExitStatus op_zero2(DisasContext *s, DisasOps *o)
3377 o->out = tcg_const_i64(0);
3383 /* ====================================================================== */
3384 /* The "Cc OUTput" generators. Given the generated output (and in some cases
3385 the original inputs), update the various cc data structures in order to
3386 be able to compute the new condition code. */
3388 static void cout_abs32(DisasContext *s, DisasOps *o)
3390 gen_op_update1_cc_i64(s, CC_OP_ABS_32, o->out);
3393 static void cout_abs64(DisasContext *s, DisasOps *o)
3395 gen_op_update1_cc_i64(s, CC_OP_ABS_64, o->out);
3398 static void cout_adds32(DisasContext *s, DisasOps *o)
3400 gen_op_update3_cc_i64(s, CC_OP_ADD_32, o->in1, o->in2, o->out);
3403 static void cout_adds64(DisasContext *s, DisasOps *o)
3405 gen_op_update3_cc_i64(s, CC_OP_ADD_64, o->in1, o->in2, o->out);
3408 static void cout_addu32(DisasContext *s, DisasOps *o)
3410 gen_op_update3_cc_i64(s, CC_OP_ADDU_32, o->in1, o->in2, o->out);
3413 static void cout_addu64(DisasContext *s, DisasOps *o)
3415 gen_op_update3_cc_i64(s, CC_OP_ADDU_64, o->in1, o->in2, o->out);
3418 static void cout_addc32(DisasContext *s, DisasOps *o)
3420 gen_op_update3_cc_i64(s, CC_OP_ADDC_32, o->in1, o->in2, o->out);
3423 static void cout_addc64(DisasContext *s, DisasOps *o)
3425 gen_op_update3_cc_i64(s, CC_OP_ADDC_64, o->in1, o->in2, o->out);
3428 static void cout_cmps32(DisasContext *s, DisasOps *o)
3430 gen_op_update2_cc_i64(s, CC_OP_LTGT_32, o->in1, o->in2);
3433 static void cout_cmps64(DisasContext *s, DisasOps *o)
3435 gen_op_update2_cc_i64(s, CC_OP_LTGT_64, o->in1, o->in2);
3438 static void cout_cmpu32(DisasContext *s, DisasOps *o)
3440 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_32, o->in1, o->in2);
3443 static void cout_cmpu64(DisasContext *s, DisasOps *o)
3445 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, o->in1, o->in2);
3448 static void cout_f32(DisasContext *s, DisasOps *o)
3450 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, o->out);
3453 static void cout_f64(DisasContext *s, DisasOps *o)
3455 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, o->out);
3458 static void cout_f128(DisasContext *s, DisasOps *o)
3460 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, o->out, o->out2);
3463 static void cout_nabs32(DisasContext *s, DisasOps *o)
3465 gen_op_update1_cc_i64(s, CC_OP_NABS_32, o->out);
3468 static void cout_nabs64(DisasContext *s, DisasOps *o)
3470 gen_op_update1_cc_i64(s, CC_OP_NABS_64, o->out);
3473 static void cout_neg32(DisasContext *s, DisasOps *o)
3475 gen_op_update1_cc_i64(s, CC_OP_COMP_32, o->out);
3478 static void cout_neg64(DisasContext *s, DisasOps *o)
3480 gen_op_update1_cc_i64(s, CC_OP_COMP_64, o->out);
3483 static void cout_nz32(DisasContext *s, DisasOps *o)
3485 tcg_gen_ext32u_i64(cc_dst, o->out);
3486 gen_op_update1_cc_i64(s, CC_OP_NZ, cc_dst);
3489 static void cout_nz64(DisasContext *s, DisasOps *o)
3491 gen_op_update1_cc_i64(s, CC_OP_NZ, o->out);
3494 static void cout_s32(DisasContext *s, DisasOps *o)
3496 gen_op_update1_cc_i64(s, CC_OP_LTGT0_32, o->out);
3499 static void cout_s64(DisasContext *s, DisasOps *o)
3501 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, o->out);
3504 static void cout_subs32(DisasContext *s, DisasOps *o)
3506 gen_op_update3_cc_i64(s, CC_OP_SUB_32, o->in1, o->in2, o->out);
3509 static void cout_subs64(DisasContext *s, DisasOps *o)
3511 gen_op_update3_cc_i64(s, CC_OP_SUB_64, o->in1, o->in2, o->out);
3514 static void cout_subu32(DisasContext *s, DisasOps *o)
3516 gen_op_update3_cc_i64(s, CC_OP_SUBU_32, o->in1, o->in2, o->out);
3519 static void cout_subu64(DisasContext *s, DisasOps *o)
3521 gen_op_update3_cc_i64(s, CC_OP_SUBU_64, o->in1, o->in2, o->out);
3524 static void cout_subb32(DisasContext *s, DisasOps *o)
3526 gen_op_update3_cc_i64(s, CC_OP_SUBB_32, o->in1, o->in2, o->out);
3529 static void cout_subb64(DisasContext *s, DisasOps *o)
3531 gen_op_update3_cc_i64(s, CC_OP_SUBB_64, o->in1, o->in2, o->out);
3534 static void cout_tm32(DisasContext *s, DisasOps *o)
3536 gen_op_update2_cc_i64(s, CC_OP_TM_32, o->in1, o->in2);
3539 static void cout_tm64(DisasContext *s, DisasOps *o)
3541 gen_op_update2_cc_i64(s, CC_OP_TM_64, o->in1, o->in2);
3544 /* ====================================================================== */
3545 /* The "PREPeration" generators. These initialize the DisasOps.OUT fields
3546 with the TCG register to which we will write. Used in combination with
3547 the "wout" generators, in some cases we need a new temporary, and in
3548 some cases we can write to a TCG global. */
3550 static void prep_new(DisasContext *s, DisasFields *f, DisasOps *o)
3552 o->out = tcg_temp_new_i64();
3555 static void prep_new_P(DisasContext *s, DisasFields *f, DisasOps *o)
3557 o->out = tcg_temp_new_i64();
3558 o->out2 = tcg_temp_new_i64();
3561 static void prep_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3563 o->out = regs[get_field(f, r1)];
3567 static void prep_r1_P(DisasContext *s, DisasFields *f, DisasOps *o)
3569 /* ??? Specification exception: r1 must be even. */
3570 int r1 = get_field(f, r1);
3572 o->out2 = regs[(r1 + 1) & 15];
3573 o->g_out = o->g_out2 = true;
3576 static void prep_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3578 o->out = fregs[get_field(f, r1)];
3582 static void prep_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3584 /* ??? Specification exception: r1 must be < 14. */
3585 int r1 = get_field(f, r1);
3587 o->out2 = fregs[(r1 + 2) & 15];
3588 o->g_out = o->g_out2 = true;
3591 /* ====================================================================== */
3592 /* The "Write OUTput" generators. These generally perform some non-trivial
3593 copy of data to TCG globals, or to main memory. The trivial cases are
3594 generally handled by having a "prep" generator install the TCG global
3595 as the destination of the operation. */
3597 static void wout_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3599 store_reg(get_field(f, r1), o->out);
3602 static void wout_r1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3604 int r1 = get_field(f, r1);
3605 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 8);
3608 static void wout_r1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3610 int r1 = get_field(f, r1);
3611 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 16);
3614 static void wout_r1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3616 store_reg32_i64(get_field(f, r1), o->out);
3619 static void wout_r1_P32(DisasContext *s, DisasFields *f, DisasOps *o)
3621 /* ??? Specification exception: r1 must be even. */
3622 int r1 = get_field(f, r1);
3623 store_reg32_i64(r1, o->out);
3624 store_reg32_i64((r1 + 1) & 15, o->out2);
3627 static void wout_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3629 /* ??? Specification exception: r1 must be even. */
3630 int r1 = get_field(f, r1);
3631 store_reg32_i64((r1 + 1) & 15, o->out);
3632 tcg_gen_shri_i64(o->out, o->out, 32);
3633 store_reg32_i64(r1, o->out);
3636 static void wout_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3638 store_freg32_i64(get_field(f, r1), o->out);
3641 static void wout_f1(DisasContext *s, DisasFields *f, DisasOps *o)
3643 store_freg(get_field(f, r1), o->out);
3646 static void wout_x1(DisasContext *s, DisasFields *f, DisasOps *o)
3648 /* ??? Specification exception: r1 must be < 14. */
3649 int f1 = get_field(s->fields, r1);
3650 store_freg(f1, o->out);
3651 store_freg((f1 + 2) & 15, o->out2);
3654 static void wout_cond_r1r2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3656 if (get_field(f, r1) != get_field(f, r2)) {
3657 store_reg32_i64(get_field(f, r1), o->out);
3661 static void wout_cond_e1e2(DisasContext *s, DisasFields *f, DisasOps *o)
3663 if (get_field(f, r1) != get_field(f, r2)) {
3664 store_freg32_i64(get_field(f, r1), o->out);
3668 static void wout_m1_8(DisasContext *s, DisasFields *f, DisasOps *o)
3670 tcg_gen_qemu_st8(o->out, o->addr1, get_mem_index(s));
3673 static void wout_m1_16(DisasContext *s, DisasFields *f, DisasOps *o)
3675 tcg_gen_qemu_st16(o->out, o->addr1, get_mem_index(s));
3678 static void wout_m1_32(DisasContext *s, DisasFields *f, DisasOps *o)
3680 tcg_gen_qemu_st32(o->out, o->addr1, get_mem_index(s));
3683 static void wout_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3685 tcg_gen_qemu_st64(o->out, o->addr1, get_mem_index(s));
3688 static void wout_m2_32(DisasContext *s, DisasFields *f, DisasOps *o)
3690 tcg_gen_qemu_st32(o->out, o->in2, get_mem_index(s));
3693 /* ====================================================================== */
3694 /* The "INput 1" generators. These load the first operand to an insn. */
3696 static void in1_r1(DisasContext *s, DisasFields *f, DisasOps *o)
3698 o->in1 = load_reg(get_field(f, r1));
3701 static void in1_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3703 o->in1 = regs[get_field(f, r1)];
3707 static void in1_r1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3709 o->in1 = tcg_temp_new_i64();
3710 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r1)]);
3713 static void in1_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3715 o->in1 = tcg_temp_new_i64();
3716 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r1)]);
3719 static void in1_r1_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
3721 o->in1 = tcg_temp_new_i64();
3722 tcg_gen_shri_i64(o->in1, regs[get_field(f, r1)], 32);
3725 static void in1_r1p1(DisasContext *s, DisasFields *f, DisasOps *o)
3727 /* ??? Specification exception: r1 must be even. */
3728 int r1 = get_field(f, r1);
3729 o->in1 = load_reg((r1 + 1) & 15);
3732 static void in1_r1p1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3734 /* ??? Specification exception: r1 must be even. */
3735 int r1 = get_field(f, r1);
3736 o->in1 = tcg_temp_new_i64();
3737 tcg_gen_ext32s_i64(o->in1, regs[(r1 + 1) & 15]);
3740 static void in1_r1p1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3742 /* ??? Specification exception: r1 must be even. */
3743 int r1 = get_field(f, r1);
3744 o->in1 = tcg_temp_new_i64();
3745 tcg_gen_ext32u_i64(o->in1, regs[(r1 + 1) & 15]);
3748 static void in1_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
3750 /* ??? Specification exception: r1 must be even. */
3751 int r1 = get_field(f, r1);
3752 o->in1 = tcg_temp_new_i64();
3753 tcg_gen_concat32_i64(o->in1, regs[r1 + 1], regs[r1]);
3756 static void in1_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3758 o->in1 = load_reg(get_field(f, r2));
3761 static void in1_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3763 o->in1 = load_reg(get_field(f, r3));
3766 static void in1_r3_o(DisasContext *s, DisasFields *f, DisasOps *o)
3768 o->in1 = regs[get_field(f, r3)];
3772 static void in1_r3_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3774 o->in1 = tcg_temp_new_i64();
3775 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r3)]);
3778 static void in1_r3_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3780 o->in1 = tcg_temp_new_i64();
3781 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r3)]);
3784 static void in1_e1(DisasContext *s, DisasFields *f, DisasOps *o)
3786 o->in1 = load_freg32_i64(get_field(f, r1));
3789 static void in1_f1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3791 o->in1 = fregs[get_field(f, r1)];
3795 static void in1_x1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3797 /* ??? Specification exception: r1 must be < 14. */
3798 int r1 = get_field(f, r1);
3800 o->out2 = fregs[(r1 + 2) & 15];
3801 o->g_out = o->g_out2 = true;
3804 static void in1_la1(DisasContext *s, DisasFields *f, DisasOps *o)
3806 o->addr1 = get_address(s, 0, get_field(f, b1), get_field(f, d1));
3809 static void in1_la2(DisasContext *s, DisasFields *f, DisasOps *o)
3811 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3812 o->addr1 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3815 static void in1_m1_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3818 o->in1 = tcg_temp_new_i64();
3819 tcg_gen_qemu_ld8u(o->in1, o->addr1, get_mem_index(s));
3822 static void in1_m1_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3825 o->in1 = tcg_temp_new_i64();
3826 tcg_gen_qemu_ld16s(o->in1, o->addr1, get_mem_index(s));
3829 static void in1_m1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3832 o->in1 = tcg_temp_new_i64();
3833 tcg_gen_qemu_ld16u(o->in1, o->addr1, get_mem_index(s));
3836 static void in1_m1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3839 o->in1 = tcg_temp_new_i64();
3840 tcg_gen_qemu_ld32s(o->in1, o->addr1, get_mem_index(s));
3843 static void in1_m1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3846 o->in1 = tcg_temp_new_i64();
3847 tcg_gen_qemu_ld32u(o->in1, o->addr1, get_mem_index(s));
3850 static void in1_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
3853 o->in1 = tcg_temp_new_i64();
3854 tcg_gen_qemu_ld64(o->in1, o->addr1, get_mem_index(s));
3857 /* ====================================================================== */
3858 /* The "INput 2" generators. These load the second operand to an insn. */
3860 static void in2_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
3862 o->in2 = regs[get_field(f, r1)];
3866 static void in2_r1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3868 o->in2 = tcg_temp_new_i64();
3869 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r1)]);
3872 static void in2_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3874 o->in2 = tcg_temp_new_i64();
3875 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r1)]);
3878 static void in2_r2(DisasContext *s, DisasFields *f, DisasOps *o)
3880 o->in2 = load_reg(get_field(f, r2));
3883 static void in2_r2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3885 o->in2 = regs[get_field(f, r2)];
3889 static void in2_r2_nz(DisasContext *s, DisasFields *f, DisasOps *o)
3891 int r2 = get_field(f, r2);
3893 o->in2 = load_reg(r2);
3897 static void in2_r2_8s(DisasContext *s, DisasFields *f, DisasOps *o)
3899 o->in2 = tcg_temp_new_i64();
3900 tcg_gen_ext8s_i64(o->in2, regs[get_field(f, r2)]);
3903 static void in2_r2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3905 o->in2 = tcg_temp_new_i64();
3906 tcg_gen_ext8u_i64(o->in2, regs[get_field(f, r2)]);
3909 static void in2_r2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3911 o->in2 = tcg_temp_new_i64();
3912 tcg_gen_ext16s_i64(o->in2, regs[get_field(f, r2)]);
3915 static void in2_r2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3917 o->in2 = tcg_temp_new_i64();
3918 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r2)]);
3921 static void in2_r3(DisasContext *s, DisasFields *f, DisasOps *o)
3923 o->in2 = load_reg(get_field(f, r3));
3926 static void in2_r2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
3928 o->in2 = tcg_temp_new_i64();
3929 tcg_gen_ext32s_i64(o->in2, regs[get_field(f, r2)]);
3932 static void in2_r2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
3934 o->in2 = tcg_temp_new_i64();
3935 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r2)]);
3938 static void in2_e2(DisasContext *s, DisasFields *f, DisasOps *o)
3940 o->in2 = load_freg32_i64(get_field(f, r2));
3943 static void in2_f2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3945 o->in2 = fregs[get_field(f, r2)];
3949 static void in2_x2_o(DisasContext *s, DisasFields *f, DisasOps *o)
3951 /* ??? Specification exception: r1 must be < 14. */
3952 int r2 = get_field(f, r2);
3954 o->in2 = fregs[(r2 + 2) & 15];
3955 o->g_in1 = o->g_in2 = true;
3958 static void in2_ra2(DisasContext *s, DisasFields *f, DisasOps *o)
3960 o->in2 = get_address(s, 0, get_field(f, r2), 0);
3963 static void in2_a2(DisasContext *s, DisasFields *f, DisasOps *o)
3965 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
3966 o->in2 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
3969 static void in2_ri2(DisasContext *s, DisasFields *f, DisasOps *o)
3971 o->in2 = tcg_const_i64(s->pc + (int64_t)get_field(f, i2) * 2);
3974 static void in2_sh32(DisasContext *s, DisasFields *f, DisasOps *o)
3976 help_l2_shift(s, f, o, 31);
3979 static void in2_sh64(DisasContext *s, DisasFields *f, DisasOps *o)
3981 help_l2_shift(s, f, o, 63);
3984 static void in2_m2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
3987 tcg_gen_qemu_ld8u(o->in2, o->in2, get_mem_index(s));
3990 static void in2_m2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
3993 tcg_gen_qemu_ld16s(o->in2, o->in2, get_mem_index(s));
3996 static void in2_m2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
3999 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
4002 static void in2_m2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4005 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
4008 static void in2_m2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4011 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
4014 static void in2_m2_64(DisasContext *s, DisasFields *f, DisasOps *o)
4017 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
4020 static void in2_mri2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4023 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
4026 static void in2_mri2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4029 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
4032 static void in2_mri2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4035 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
4038 static void in2_mri2_64(DisasContext *s, DisasFields *f, DisasOps *o)
4041 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
4044 static void in2_i2(DisasContext *s, DisasFields *f, DisasOps *o)
4046 o->in2 = tcg_const_i64(get_field(f, i2));
4049 static void in2_i2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
4051 o->in2 = tcg_const_i64((uint8_t)get_field(f, i2));
4054 static void in2_i2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4056 o->in2 = tcg_const_i64((uint16_t)get_field(f, i2));
4059 static void in2_i2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4061 o->in2 = tcg_const_i64((uint32_t)get_field(f, i2));
4064 static void in2_i2_16u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
4066 uint64_t i2 = (uint16_t)get_field(f, i2);
4067 o->in2 = tcg_const_i64(i2 << s->insn->data);
4070 static void in2_i2_32u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
4072 uint64_t i2 = (uint32_t)get_field(f, i2);
4073 o->in2 = tcg_const_i64(i2 << s->insn->data);
4076 /* ====================================================================== */
4078 /* Find opc within the table of insns. This is formulated as a switch
4079 statement so that (1) we get compile-time notice of cut-paste errors
4080 for duplicated opcodes, and (2) the compiler generates the binary
4081 search tree, rather than us having to post-process the table. */
4083 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
4084 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
4086 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
4088 enum DisasInsnEnum {
4089 #include "insn-data.def"
4093 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
4098 .help_in1 = in1_##I1, \
4099 .help_in2 = in2_##I2, \
4100 .help_prep = prep_##P, \
4101 .help_wout = wout_##W, \
4102 .help_cout = cout_##CC, \
4103 .help_op = op_##OP, \
4107 /* Allow 0 to be used for NULL in the table below. */
4115 static const DisasInsn insn_info[] = {
4116 #include "insn-data.def"
4120 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
4121 case OPC: return &insn_info[insn_ ## NM];
4123 static const DisasInsn *lookup_opc(uint16_t opc)
4126 #include "insn-data.def"
4135 /* Extract a field from the insn. The INSN should be left-aligned in
4136 the uint64_t so that we can more easily utilize the big-bit-endian
4137 definitions we extract from the Principals of Operation. */
4139 static void extract_field(DisasFields *o, const DisasField *f, uint64_t insn)
4147 /* Zero extract the field from the insn. */
4148 r = (insn << f->beg) >> (64 - f->size);
4150 /* Sign-extend, or un-swap the field as necessary. */
4152 case 0: /* unsigned */
4154 case 1: /* signed */
4155 assert(f->size <= 32);
4156 m = 1u << (f->size - 1);
4159 case 2: /* dl+dh split, signed 20 bit. */
4160 r = ((int8_t)r << 12) | (r >> 8);
4166 /* Validate that the "compressed" encoding we selected above is valid.
4167 I.e. we havn't make two different original fields overlap. */
4168 assert(((o->presentC >> f->indexC) & 1) == 0);
4169 o->presentC |= 1 << f->indexC;
4170 o->presentO |= 1 << f->indexO;
4172 o->c[f->indexC] = r;
4175 /* Lookup the insn at the current PC, extracting the operands into O and
4176 returning the info struct for the insn. Returns NULL for invalid insn. */
4178 static const DisasInsn *extract_insn(CPUS390XState *env, DisasContext *s,
4181 uint64_t insn, pc = s->pc;
4183 const DisasInsn *info;
4185 insn = ld_code2(env, pc);
4186 op = (insn >> 8) & 0xff;
4187 ilen = get_ilen(op);
4188 s->next_pc = s->pc + ilen;
4195 insn = ld_code4(env, pc) << 32;
4198 insn = (insn << 48) | (ld_code4(env, pc + 2) << 16);
4204 /* We can't actually determine the insn format until we've looked up
4205 the full insn opcode. Which we can't do without locating the
4206 secondary opcode. Assume by default that OP2 is at bit 40; for
4207 those smaller insns that don't actually have a secondary opcode
4208 this will correctly result in OP2 = 0. */
4214 case 0xb2: /* S, RRF, RRE */
4215 case 0xb3: /* RRE, RRD, RRF */
4216 case 0xb9: /* RRE, RRF */
4217 case 0xe5: /* SSE, SIL */
4218 op2 = (insn << 8) >> 56;
4222 case 0xc0: /* RIL */
4223 case 0xc2: /* RIL */
4224 case 0xc4: /* RIL */
4225 case 0xc6: /* RIL */
4226 case 0xc8: /* SSF */
4227 case 0xcc: /* RIL */
4228 op2 = (insn << 12) >> 60;
4230 case 0xd0 ... 0xdf: /* SS */
4236 case 0xee ... 0xf3: /* SS */
4237 case 0xf8 ... 0xfd: /* SS */
4241 op2 = (insn << 40) >> 56;
4245 memset(f, 0, sizeof(*f));
4249 /* Lookup the instruction. */
4250 info = lookup_opc(op << 8 | op2);
4252 /* If we found it, extract the operands. */
4254 DisasFormat fmt = info->fmt;
4257 for (i = 0; i < NUM_C_FIELD; ++i) {
4258 extract_field(f, &format_info[fmt].op[i], insn);
4264 static ExitStatus translate_one(CPUS390XState *env, DisasContext *s)
4266 const DisasInsn *insn;
4267 ExitStatus ret = NO_EXIT;
4271 /* Search for the insn in the table. */
4272 insn = extract_insn(env, s, &f);
4274 /* Not found means unimplemented/illegal opcode. */
4276 qemu_log_mask(LOG_UNIMP, "unimplemented opcode 0x%02x%02x\n",
4278 gen_illegal_opcode(s);
4279 return EXIT_NORETURN;
4282 /* Set up the strutures we use to communicate with the helpers. */
4285 o.g_out = o.g_out2 = o.g_in1 = o.g_in2 = false;
4286 TCGV_UNUSED_I64(o.out);
4287 TCGV_UNUSED_I64(o.out2);
4288 TCGV_UNUSED_I64(o.in1);
4289 TCGV_UNUSED_I64(o.in2);
4290 TCGV_UNUSED_I64(o.addr1);
4292 /* Implement the instruction. */
4293 if (insn->help_in1) {
4294 insn->help_in1(s, &f, &o);
4296 if (insn->help_in2) {
4297 insn->help_in2(s, &f, &o);
4299 if (insn->help_prep) {
4300 insn->help_prep(s, &f, &o);
4302 if (insn->help_op) {
4303 ret = insn->help_op(s, &o);
4305 if (insn->help_wout) {
4306 insn->help_wout(s, &f, &o);
4308 if (insn->help_cout) {
4309 insn->help_cout(s, &o);
4312 /* Free any temporaries created by the helpers. */
4313 if (!TCGV_IS_UNUSED_I64(o.out) && !o.g_out) {
4314 tcg_temp_free_i64(o.out);
4316 if (!TCGV_IS_UNUSED_I64(o.out2) && !o.g_out2) {
4317 tcg_temp_free_i64(o.out2);
4319 if (!TCGV_IS_UNUSED_I64(o.in1) && !o.g_in1) {
4320 tcg_temp_free_i64(o.in1);
4322 if (!TCGV_IS_UNUSED_I64(o.in2) && !o.g_in2) {
4323 tcg_temp_free_i64(o.in2);
4325 if (!TCGV_IS_UNUSED_I64(o.addr1)) {
4326 tcg_temp_free_i64(o.addr1);
4329 /* Advance to the next instruction. */
4334 static inline void gen_intermediate_code_internal(CPUS390XState *env,
4335 TranslationBlock *tb,
4339 target_ulong pc_start;
4340 uint64_t next_page_start;
4341 uint16_t *gen_opc_end;
4343 int num_insns, max_insns;
4351 if (!(tb->flags & FLAG_MASK_64)) {
4352 pc_start &= 0x7fffffff;
4357 dc.cc_op = CC_OP_DYNAMIC;
4358 do_debug = dc.singlestep_enabled = env->singlestep_enabled;
4360 gen_opc_end = tcg_ctx.gen_opc_buf + OPC_MAX_SIZE;
4362 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
4365 max_insns = tb->cflags & CF_COUNT_MASK;
4366 if (max_insns == 0) {
4367 max_insns = CF_COUNT_MASK;
4374 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4378 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4381 tcg_ctx.gen_opc_pc[lj] = dc.pc;
4382 gen_opc_cc_op[lj] = dc.cc_op;
4383 tcg_ctx.gen_opc_instr_start[lj] = 1;
4384 tcg_ctx.gen_opc_icount[lj] = num_insns;
4386 if (++num_insns == max_insns && (tb->cflags & CF_LAST_IO)) {
4390 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) {
4391 tcg_gen_debug_insn_start(dc.pc);
4395 if (unlikely(!QTAILQ_EMPTY(&env->breakpoints))) {
4396 QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
4397 if (bp->pc == dc.pc) {
4398 status = EXIT_PC_STALE;
4404 if (status == NO_EXIT) {
4405 status = translate_one(env, &dc);
4408 /* If we reach a page boundary, are single stepping,
4409 or exhaust instruction count, stop generation. */
4410 if (status == NO_EXIT
4411 && (dc.pc >= next_page_start
4412 || tcg_ctx.gen_opc_ptr >= gen_opc_end
4413 || num_insns >= max_insns
4415 || env->singlestep_enabled)) {
4416 status = EXIT_PC_STALE;
4418 } while (status == NO_EXIT);
4420 if (tb->cflags & CF_LAST_IO) {
4429 update_psw_addr(&dc);
4431 case EXIT_PC_UPDATED:
4432 /* Next TB starts off with CC_OP_DYNAMIC, so make sure the
4433 cc op type is in env */
4435 /* Exit the TB, either by raising a debug exception or by return. */
4437 gen_exception(EXCP_DEBUG);
4446 gen_icount_end(tb, num_insns);
4447 *tcg_ctx.gen_opc_ptr = INDEX_op_end;
4449 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
4452 tcg_ctx.gen_opc_instr_start[lj++] = 0;
4455 tb->size = dc.pc - pc_start;
4456 tb->icount = num_insns;
4459 #if defined(S390X_DEBUG_DISAS)
4460 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
4461 qemu_log("IN: %s\n", lookup_symbol(pc_start));
4462 log_target_disas(env, pc_start, dc.pc - pc_start, 1);
4468 void gen_intermediate_code (CPUS390XState *env, struct TranslationBlock *tb)
4470 gen_intermediate_code_internal(env, tb, 0);
4473 void gen_intermediate_code_pc (CPUS390XState *env, struct TranslationBlock *tb)
4475 gen_intermediate_code_internal(env, tb, 1);
4478 void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb, int pc_pos)
4481 env->psw.addr = tcg_ctx.gen_opc_pc[pc_pos];
4482 cc_op = gen_opc_cc_op[pc_pos];
4483 if ((cc_op != CC_OP_DYNAMIC) && (cc_op != CC_OP_STATIC)) {