any.bitfield.cpuavx512vl = 0;
}
}
+
+ /* Dual non-APX/APX templates need massaging from what APX_F() in the
+ opcode table has produced. While the direct transformation of the
+ incoming cpuid&(cpuid|APX_F) would be to cpuid&(cpuid) / cpuid&(APX_F)
+ respectively, it's cheaper to move to just cpuid / cpuid&APX_F
+ instead. */
+ if (any.bitfield.cpuapx_f
+ && (any.bitfield.cpubmi || any.bitfield.cpubmi2
+ || any.bitfield.cpuavx512f || any.bitfield.cpuavx512bw
+ || any.bitfield.cpuavx512dq || any.bitfield.cpuamx_tile
+ || any.bitfield.cpucmpccxadd))
+ {
+ /* These checks (verifying that APX_F() was properly used in the
+ opcode table entry) make sure there's no need for an "else" to
+ the "if()" below. */
+ gas_assert (!cpu_flags_all_zero (&all));
+ cpu = cpu_flags_and (all, any);
+ gas_assert (cpu_flags_equal (&cpu, &all));
+
+ if (need_evex_encoding (t))
+ all = any;
+
+ memset (&any, 0, sizeof (any));
+ }
}
if (flag_code != CODE_64BIT)
.*:25: Error: `andn' is not supported on `x86_64.nobmi'
.*:28: Error: `bzhi' is not supported on `x86_64.nobmi2'
.*:29: Error: `bzhi' is not supported on `x86_64.nobmi2'
+.*:33: Error: .*`andn'.*
+.*:34: Error: .*`bzhi'.*
+.*:35: Error: .*`kmovw'.*
+.*:36: Error: .*`kmovq'.*
+.*:37: Error: .*`kmovb'.*
+.*:38: Error: .*`ldtilecfg'.*
+.*:39: Error: .*`cmpexadd'.*
GAS LISTING .*
#...
[ ]*1[ ]+\# Check illegal 64bit APX EVEX promoted instructions
.arch .nobmi2
bzhi %r16,%r15,%r11
bzhi %r15,%r15,%r11
+
+ .arch default
+ .arch .noapx_f
+ {evex} andn %r15, %r15, %r11
+ {evex} bzhi %r15, %r15, %r11
+ {evex} kmovw %k1, %r8d
+ {evex} kmovq %k1, %r8
+ {evex} kmovb %k1, %r8d
+ {evex} ldtilecfg (%r8)
+ {evex} cmpexadd %rax, %rcx, (%r8)
#define DstVVVV VexVVVV=VexVVVV_DST
-// The template supports VEX format for cpuid and EVEX format for cpuid & apx_f.
+// The template supports VEX format for cpuid and EVEX format for cpuid & APX_F.
+// While therefore we really mean cpuid|(cpuid&APX_F) here, this can't be
+// expressed in the generated templates. It's equivalent to just cpuid|APX_F
+// anyway, but that is not what we want (as APX_F alone isn't a sufficient
+// prereq for such insns). Instead the assembler will massage the CPU specifier
+// to the equivalent of either cpuid&(cpuid) or cpuid&(APX_F) (or something
+// substantially similar), depending on what encoding was requested.
#define APX_F(cpuid) cpuid&(cpuid|APX_F)
// The EVEX purpose of StaticRounding appears only together with SAE. Re-use