#ifdef __GNUC__
# ifdef __ICC
-# define MC(x) M64 (c.mmx_ ## x)
+# define MC(x) to_m64 (c.mmx_ ## x)
# else
# define MC(x) ((__m64)c.mmx_ ## x)
# endif
#endif
static force_inline __m64
-M64 (uint64_t x)
+to_m64 (uint64_t x)
{
#ifdef __ICC
return _mm_cvtsi64_m64 (x);
}
static force_inline uint64_t
-UINT64 (__m64 x)
+to_uint64 (__m64 x)
{
#ifdef __ICC
return _mm_cvtm64_si64 (x);
while (w && (unsigned long)dst & 7)
{
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (over (vsrc, vsrca, vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
while (w)
{
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (over (vsrc, vsrca, vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
{
__m64 vsrc = load8888 (*src);
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (
over (vsrc, expand_alpha (vsrc), vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
{
__m64 vsrc = load8888 (*src);
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (over (vsrc, expand_alpha (vsrc), vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
if (m)
{
__m64 vdest = in_over (vsrc, vsrca,
- expand_alpha_rev (M64 (m)),
+ expand_alpha_rev (to_m64 (m)),
load8888 (*dst));
*dst = store8888 (vdest);
vdest = *(__m64 *)dst;
- dest0 = in_over (vsrc, vsrca, expand_alpha_rev (M64 (m0)),
+ dest0 = in_over (vsrc, vsrca, expand_alpha_rev (to_m64 (m0)),
expand8888 (vdest, 0));
- dest1 = in_over (vsrc, vsrca, expand_alpha_rev (M64 (m1)),
+ dest1 = in_over (vsrc, vsrca, expand_alpha_rev (to_m64 (m1)),
expand8888 (vdest, 1));
*(__m64 *)dst = pack8888 (dest0, dest1);
__m64 vdest = load8888 (*dst);
vdest = in_over (
- vsrc, vsrca, expand_alpha_rev (M64 (m)), vdest);
+ vsrc, vsrca, expand_alpha_rev (to_m64 (m)), vdest);
*dst = store8888 (vdest);
}
}
fill = ((uint64_t)xor << 32) | xor;
- vfill = M64 (fill);
+ vfill = to_m64 (fill);
#ifdef __GNUC__
__asm__ (
if (m)
{
- __m64 vdest = in (vsrc, expand_alpha_rev (M64 (m)));
+ __m64 vdest = in (vsrc, expand_alpha_rev (to_m64 (m)));
*dst = store8888 (vdest);
}
vdest = *(__m64 *)dst;
- dest0 = in (vsrc, expand_alpha_rev (M64 (m0)));
- dest1 = in (vsrc, expand_alpha_rev (M64 (m1)));
+ dest0 = in (vsrc, expand_alpha_rev (to_m64 (m0)));
+ dest1 = in (vsrc, expand_alpha_rev (to_m64 (m1)));
*(__m64 *)dst = pack8888 (dest0, dest1);
}
{
__m64 vdest = load8888 (*dst);
- vdest = in (vsrc, expand_alpha_rev (M64 (m)));
+ vdest = in (vsrc, expand_alpha_rev (to_m64 (m)));
*dst = store8888 (vdest);
}
else
vsrca = expand_alpha (vsrc);
tmp = pack_565 (vsrc, _mm_setzero_si64 (), 0);
- src16 = UINT64 (tmp);
+ src16 = to_uint64 (tmp);
srcsrcsrcsrc =
(uint64_t)src16 << 48 | (uint64_t)src16 << 32 |
if (m)
{
uint64_t d = *dst;
- __m64 vd = M64 (d);
+ __m64 vd = to_m64 (d);
__m64 vdest = in_over (
- vsrc, vsrca, expand_alpha_rev (M64 (m)), expand565 (vd, 0));
+ vsrc, vsrca, expand_alpha_rev (to_m64 (m)), expand565 (vd, 0));
vd = pack_565 (vdest, _mm_setzero_si64 (), 0);
- *dst = UINT64 (vd);
+ *dst = to_uint64 (vd);
}
w--;
vdest = *(__m64 *)dst;
- vm0 = M64 (m0);
+ vm0 = to_m64 (m0);
vdest = pack_565 (in_over (vsrc, vsrca, expand_alpha_rev (vm0),
expand565 (vdest, 0)), vdest, 0);
- vm1 = M64 (m1);
+ vm1 = to_m64 (m1);
vdest = pack_565 (in_over (vsrc, vsrca, expand_alpha_rev (vm1),
expand565 (vdest, 1)), vdest, 1);
- vm2 = M64 (m2);
+ vm2 = to_m64 (m2);
vdest = pack_565 (in_over (vsrc, vsrca, expand_alpha_rev (vm2),
expand565 (vdest, 2)), vdest, 2);
- vm3 = M64 (m3);
+ vm3 = to_m64 (m3);
vdest = pack_565 (in_over (vsrc, vsrca, expand_alpha_rev (vm3),
expand565 (vdest, 3)), vdest, 3);
if (m)
{
uint64_t d = *dst;
- __m64 vd = M64 (d);
- __m64 vdest = in_over (vsrc, vsrca, expand_alpha_rev (M64 (m)),
+ __m64 vd = to_m64 (d);
+ __m64 vdest = in_over (vsrc, vsrca, expand_alpha_rev (to_m64 (m)),
expand565 (vd, 0));
vd = pack_565 (vdest, _mm_setzero_si64 (), 0);
- *dst = UINT64 (vd);
+ *dst = to_uint64 (vd);
}
w--;
{
__m64 vsrc = load8888 (*src);
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (over_rev_non_pre (vsrc, vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
{
__m64 vsrc = load8888 (*src);
uint64_t d = *dst;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (over_rev_non_pre (vsrc, vdest), vdest, 0);
- *dst = UINT64 (vdest);
+ *dst = to_uint64 (vdest);
w--;
dst++;
if (m)
{
uint64_t d = *q;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (in_over (vsrc, vsrca, load8888 (m), vdest), vdest, 0);
- *q = UINT64 (vdest);
+ *q = to_uint64 (vdest);
}
twidth--;
if (m)
{
uint64_t d = *q;
- __m64 vdest = expand565 (M64 (d), 0);
+ __m64 vdest = expand565 (to_m64 (d), 0);
vdest = pack_565 (in_over (vsrc, vsrca, load8888 (m), vdest), vdest, 0);
- *q = UINT64 (vdest);
+ *q = to_uint64 (vdest);
}
twidth--;
while (w >= 2)
{
dst64 = _mm_adds_pu8 (*(__m64*)src, *(__m64*)dst);
- *(uint64_t*)dst = UINT64 (dst64);
+ *(uint64_t*)dst = to_uint64 (dst64);
dst += 2;
src += 2;
w -= 2;
else
{
__m64 sa = expand_alpha (s);
- __m64 vm = expand_alpha_rev (M64 (m));
+ __m64 vm = expand_alpha_rev (to_m64 (m));
__m64 vdest = in_over (s, sa, vm, load8888 (*dst));
*dst = store8888 (vdest);