* i386.c (classify_argument): Properly classify SSE/MMX modes and VOIDmode.
(construct_container): Fix handling of SSE operands.
(ix86_expand_builtin): Fix handling of 64bit pointers.
(mmx_maskmovq_rex): New pattern.
From-SVN: r49840
+Mon Feb 18 15:07:35 CET 2002 Jan Hubicka <jh@suse.cz>
+
+ * i386.c (classify_argument): Properly classify SSE/MMX modes and VOIDmode.
+ (construct_container): Fix handling of SSE operands.
+ (ix86_expand_builtin): Fix handling of 64bit pointers.
+ (mmx_maskmovq_rex): New pattern.
+
Mon Feb 18 11:55:55 CET 2002 Jan Hubicka <jh@suse.cz>
* regrename.c (kill_set_value): Handle subregs properly.
case SCmode:
classes[0] = X86_64_SSE_CLASS;
return 1;
+ case V4SFmode:
+ case V4SImode:
+ classes[0] = X86_64_SSE_CLASS;
+ classes[1] = X86_64_SSEUP_CLASS;
+ return 2;
+ case V2SFmode:
+ case V2SImode:
+ case V4HImode:
+ case V8QImode:
+ classes[0] = X86_64_SSE_CLASS;
+ return 1;
case BLKmode:
+ case VOIDmode:
return 0;
default:
abort ();
abort ();
}
if (n == 2 && class[0] == X86_64_SSE_CLASS && class[1] == X86_64_SSEUP_CLASS)
- return gen_rtx_REG (TImode, SSE_REGNO (sse_regno));
+ return gen_rtx_REG (mode, SSE_REGNO (sse_regno));
if (n == 2
&& class[0] == X86_64_X87_CLASS && class[1] == X86_64_X87UP_CLASS)
return gen_rtx_REG (TFmode, FIRST_STACK_REG);
return target;
case IX86_BUILTIN_MASKMOVQ:
- icode = CODE_FOR_mmx_maskmovq;
+ icode = TARGET_64BIT ? CODE_FOR_mmx_maskmovq_rex : CODE_FOR_mmx_maskmovq;
/* Note the arg order is different from the operand order. */
arg1 = TREE_VALUE (arglist);
arg2 = TREE_VALUE (TREE_CHAIN (arglist));
[(set (mem:V8QI (match_operand:SI 0 "register_operand" "D"))
(unspec:V8QI [(match_operand:V8QI 1 "register_operand" "y")
(match_operand:V8QI 2 "register_operand" "y")] 32))]
- "TARGET_SSE || TARGET_3DNOW_A"
+ "(TARGET_SSE || TARGET_3DNOW_A) && !TARGET_64BIT"
+ ;; @@@ check ordering of operands in intel/nonintel syntax
+ "maskmovq\t{%2, %1|%1, %2}"
+ [(set_attr "type" "sse")])
+
+(define_insn "mmx_maskmovq_rex"
+ [(set (mem:V8QI (match_operand:DI 0 "register_operand" "D"))
+ (unspec:V8QI [(match_operand:V8QI 1 "register_operand" "y")
+ (match_operand:V8QI 2 "register_operand" "y")] 32))]
+ "(TARGET_SSE || TARGET_3DNOW_A) && TARGET_64BIT"
;; @@@ check ordering of operands in intel/nonintel syntax
"maskmovq\t{%2, %1|%1, %2}"
[(set_attr "type" "sse")])