extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__))
_mm_mulhi_pu16 (__m64 __A, __m64 __B)
{
- asm("pmulhuw %1, %0\n\t"
+ asm ("pmulhuw %1, %0\n\t"
: "+y" (__A)
: "y" (__B)
);
{
__m64 ret;
- asm("pshufw %2, %1, %0\n\t"
+ asm ("pshufw %2, %1, %0\n\t"
: "=y" (ret)
: "y" (__A), "K" (__N)
);
static force_inline __m64
expand_alpha (__m64 pixel)
{
- return _mm_shuffle_pi16(pixel, _MM_SHUFFLE (3, 3, 3, 3));
+ return _mm_shuffle_pi16 (pixel, _MM_SHUFFLE (3, 3, 3, 3));
}
static force_inline __m64
expand_alpha_rev (__m64 pixel)
{
- return _mm_shuffle_pi16(pixel, _MM_SHUFFLE (0, 0, 0, 0));
+ return _mm_shuffle_pi16 (pixel, _MM_SHUFFLE (0, 0, 0, 0));
}
static force_inline __m64
invert_colors (__m64 pixel)
{
- return _mm_shuffle_pi16(pixel, _MM_SHUFFLE (3, 0, 1, 2));
+ return _mm_shuffle_pi16 (pixel, _MM_SHUFFLE (3, 0, 1, 2));
}
static force_inline __m64
return pix_multiply (src, mask);
}
-static force_inline __m64
-in_over_full_src_alpha (__m64 src, __m64 mask, __m64 dest)
-{
- src = _mm_or_si64 (src, MC (full_alpha));
-
- return over (in (src, mask), mask, dest);
-}
-
#ifndef _MSC_VER
static force_inline __m64
in_over (__m64 src, __m64 srca, __m64 mask, __m64 dest)