if (mask->pixman.im)
pixman_image_unref(mask->pixman.im);
- if (mask->cache_entry.flags.alpha)
- {
- mask->pixman.im = pixman_image_create_bits(PIXMAN_a8r8g8b8, w, h,
- (uint32_t *)mask->mask.mask,
- w * 4);
- }
- else
- {
- mask->pixman.im = pixman_image_create_bits(PIXMAN_x8r8g8b8, w, h,
- (uint32_t *)mask->mask.mask,
- w * 4);
- }
+ mask->pixman.im = pixman_image_create_bits(PIXMAN_a8r8g8b8, w, h,
+ (uint32_t *)mask->mask.mask,
+ w * 4);
#endif
}
RGBA_Image *im = (RGBA_Image *)ie;
if (im->pixman.im) pixman_image_unref(im->pixman.im);
- if (im->cache_entry.flags.alpha)
- {
- im->pixman.im = pixman_image_create_bits
- (
-// FIXME: endianess determines this
- PIXMAN_a8r8g8b8,
-// PIXMAN_b8g8r8a8,
- im->cache_entry.w, im->cache_entry.h,
- im->image.data,
- im->cache_entry.w * 4
- );
- }
- else
- {
- im->pixman.im = pixman_image_create_bits
- (
-// FIXME: endianess determines this
- PIXMAN_x8r8g8b8,
-// PIXMAN_b8g8r8x8,
- im->cache_entry.w, im->cache_entry.h,
- im->image.data,
- im->cache_entry.w * 4
- );
- }
+
+ im->pixman.im = pixman_image_create_bits
+ (
+ PIXMAN_a8r8g8b8,
+ im->cache_entry.w, im->cache_entry.h,
+ im->image.data,
+ im->cache_entry.w * 4
+ );
# else
(void)ie;
# endif
((dc->render_op == _EVAS_RENDER_COPY) ||
(dc->render_op == _EVAS_RENDER_BLEND)))
{
- pixman_op_t op = PIXMAN_OP_SRC; // _EVAS_RENDER_COPY
- if (dc->render_op == _EVAS_RENDER_BLEND)
- op = PIXMAN_OP_OVER;
+ pixman_op_t op = PIXMAN_OP_OVER; // _EVAS_RENDER_BLEND
+ if ((dc->render_op == _EVAS_RENDER_COPY) || (!src->cache_entry.flags.alpha))
+ op = PIXMAN_OP_SRC;
pixman_image_composite(op,
src->pixman.im, NULL,
(dc->render_op == _EVAS_RENDER_BLEND)))
{
// In case of pixel and color operation.
- pixman_op_t op = PIXMAN_OP_SRC; // _EVAS_RENDER_COPY
- if (dc->render_op == _EVAS_RENDER_BLEND)
- op = PIXMAN_OP_OVER;
+ pixman_op_t op = PIXMAN_OP_OVER; // _EVAS_RENDER_BLEND
+ if ((dc->render_op == _EVAS_RENDER_COPY) || (!src->cache_entry.flags.alpha))
+ op = PIXMAN_OP_SRC;
pixman_image_composite(op,
src->pixman.im, dc->mask.mask->pixman.im,